• Eager Eagle@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    2
    ·
    2 hours ago

    Waste of energy. It’s like asking a person to estimate a non-trivial angle. Either use a model trained for that task, or don’t bother.

  • Buffalox@lemmy.world
    link
    fedilink
    English
    arrow-up
    37
    arrow-down
    5
    ·
    4 hours ago

    It’s the same photo, the same model, the same question. But you won’t get the same answer. Not even close — and the differences are large enough to cause a hypoglycaemic emergency.

    OK I wonder if there’s something wrong with the photo.
    The photo:

    WTF!!??
    That’s like estimating the carbs in 2 slices of standard sandwich bread! Of course not all bread has the same amount of sugar, but a reasonable range based on an average should be a dead easy answer.

    I thought the headline sounded crazy, but try to read the article, and it actually becomes worse. I have said it many times before, these AI chatbots should not be legal, they put lives at risk.

    • inari@piefed.zip
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      2
      ·
      3 hours ago

      To be fair there’s no way of knowing what the filling is, so the AI may be guessing based on that too

      • Ludicrous0251@piefed.zip
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        1
        ·
        1 hour ago

        Friendly reminder that LLMs don’t do math, they guess what number should come next, just like words.

        It can probably link the image to the words “a photo of a sandwich on a plate”, and interpret the question as “how many calories are in a sandwich” but from there it is just guessing at the syntax of an answer, but not at finding any truth.

        It knows sandwiches have calories and those tend to be 3-4 digit numbers, but also all numbers kinda look the same, so what’s to say it’s not 2, 5, or 12 digits?

        • monkeyslikebananas2@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          24 minutes ago

          Tool-powered agents can do math though. The issue is the fuzziness of it trying to guess carbs. It doesn’t know weight, ingredients, or anything other than a picture. These tools can be useful but not for this. Maybe one day but not yet.

          Whoever claims an AI (LLM or agents) can do that and charging their users is lying and defrauding them.

      • Carnelian@lemmy.world
        link
        fedilink
        English
        arrow-up
        25
        ·
        3 hours ago

        The apps are advertising that they can do this tho. Many of them are aggressively sponsoring YouTubers who advertise you can basically just wave your phone over the food and it takes away all the “work” from traditional calorie counting apps

        • inari@piefed.zip
          link
          fedilink
          English
          arrow-up
          6
          ·
          edit-2
          3 hours ago

          That’s true, it should ask follow-up questions, or at least clarify its assumptions

    • MagicShel@lemmy.zip
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      11
      ·
      3 hours ago

      They put lives at risk the same way every single product at your local home improvement store does. When you misuse a tool for a purpose it wasn’t intended and isn’t good at, you’re going to get bad results.

      This is an issue for the educational system, not the legal system.

      • Steve@startrek.website
        link
        fedilink
        English
        arrow-up
        19
        ·
        2 hours ago

        What if the packaging on every tool at home depot grossly misrepresented its capabilities and/or purpose?

        This chainsaw cures cancer? Hot damn somebody call RFK!

        Concrete mix goes great with pancakes, etc.

        • MagicShel@lemmy.zip
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          6
          ·
          2 hours ago

          Does OpenAI claim ChatGPT is fit for those purposes? No.

          The concrete itself will happily mix into your pancakes.

          • Steve@startrek.website
            link
            fedilink
            English
            arrow-up
            12
            ·
            2 hours ago

            I think the whole point of this discussion is that the various peddlers of AI in fact do make wild claims about their capability.

            • MagicShel@lemmy.zip
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              1
              ·
              1 hour ago

              My observation is that largely it’s the downstream AI consumers who repackage it irresponsibly. That said, I don’t hang on the words of Sam Altman and it’s certain they are pushing the idea that AI is more capable than it is, but mostly what I see is them saying they built this thing and it does neat stuff and it can probably do neat stuff for you, use your imagination.

              I believe a lot of the folks developing these tools would be horrified at the irresponsible ways vendors and end users are using it.

              • XLE@piefed.social
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 minutes ago

                Sam Altman is taken seriously by the media. He is responsible for misrepresenting the product he sells. If you’re going to sling blame around, then you had better observe the words of Sam Altman.

                The thing that I think will be most impactful on that five to ten year timeframe is AI will actually discover new science.

                This sick man is taken seriously in mainstream media and politics, and it’s no exaggeration to say he has blood on his hands.

      • HuudaHarkiten@piefed.social
        link
        fedilink
        English
        arrow-up
        11
        ·
        edit-2
        2 hours ago

        As others have pointed out, this is also a problem with how they are advertising it.

        If duct tape was advertised as something that you can use to hold your roof beams together, you’d have a issue with that.

        • dream_weasel@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 hour ago

          And at the same time I wouldn’t say “hey fuck that, duct tape is terrible! It doesn’t hold beams together, I can’t use it to tow a trailer, it’s all just pretending to stick paper together because really every sliver of duct tape just sticks to the previous piece, etc etc” But that’s the cool thing we do on Lemmy.

          The ad is bad, duct tape ain’t bad.

        • MagicShel@lemmy.zip
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          2 hours ago

          I have not seen OpenAI advertise ChatGPT as capable of medical diagnosis or therapy or anything like that. If you want therapy, and you can’t afford better — because I think we can agree that AI is terrible at it, then there should be a therapy app with explicit safety controls.

          The problem is someone created a screwdriver which is handy for lots of screwdriver shaped purposes and someone is trying to carve a ham.

  • psycho_driver@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    edit-2
    3 hours ago

    Bruh a couple of months ago I asked it (Gemini) to check the number of characters, including spaces, in a potential game character name because I was working at the time and couldn’t stop to check my in-head count. It told me 21–I had counted 20. I thought I must have gotten distracted and miscounted. Later when I had time to actually focus on the issue it turned out AI had miscounted a 20 character string (maybe counting the null terminating character?).

    • boonhet@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      15
      ·
      3 hours ago

      AI doesn’t see individual characters, it sees tokens, with most tokens being a word or part of a word. That’s why per-character questions have such a high failure rate.

      • PunnyName@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        9
        ·
        2 hours ago

        It’s it doesn’t understand the simple concept of the number of letters and spaces, it needs to be reprogrammed.

        • boonhet@sopuli.xyz
          link
          fedilink
          English
          arrow-up
          14
          ·
          2 hours ago

          It doesn’t understand anything though? It never will. It’s a probability machine. If you choose to believe its output, that’s on you. I use it as a coding assistant to get boring things done faster. Fire a prompt at claude code, grab a coffee, check out the diff. But that last step is crucial. Can’t trust AI output blindly.

          • dream_weasel@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            1 hour ago

            The embedding layer post tokenization is not just a probability machine the way you’re suggesting it. You can argue that it is probabilistic with inferred sentiment, but too many people think it works like how text prediction on your phone does and that is just factually inaccurate.

            Verify output of course, but saying “it doesn’t understand anything” and “probability machine” is a borderline erroneous short sell. At the level of tokens it “understands” relationships, and those relationships are not probabilistic, though they are fundamentally approximated based on a training corpus.

            • hesh@quokk.au
              link
              fedilink
              English
              arrow-up
              2
              ·
              53 minutes ago

              Can you explain how it’s more than probability? It’s using a neural network to guess the most likely next token, isn’t it?

              • Canigou@jlai.lu
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                27 minutes ago

                You could also say that it chooses what will be the next word it will say to you. It has a few words to choose from, which it has selected in relation to the previously spoken words, your question and previous interactions (the context). The probability you’re talking about (a number) could also be seen as it’s preference among those words. I’m not sure the probability vocabulary/analogy is necessarily the best one. The best might be to not employ any analogy at all, but then you have to dig deeper into the subject to form yourself an informed opinion. This series of videos explains it better than I do : https://www.youtube.com/watch?v=aircAruvnKk&list=PLZHQObOWTQDNU6R1_67000Dx_ZCJB-3pi

  • MightEnlightenYou@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    5
    ·
    edit-2
    3 hours ago

    People should read the top comments on Hackernews instead of anyone here, they’re more informed on the topic than Lemmy is

    • Oisteink@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      3
      ·
      3 hours ago

      Yeah - if you’re after AI fanbois you should head over there. They’re not that bright, but if you check show and tell you can see what claude’s been ut to last two days