• arglebargle@lemm.ee
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    4
    ·
    1 month ago

    Yeah kinda tired of it. We don’t even have AI yet, and here people are throwing around the term right and left and then accusing everything under the sun to be generated by it.

    • 0laura@lemmy.world
      link
      fedilink
      English
      arrow-up
      18
      arrow-down
      5
      ·
      1 month ago

      ai isn’t magic, we’ve had ai for a looong time. AGI that surpasses humans? not yet.

      • arglebargle@lemm.ee
        link
        fedilink
        English
        arrow-up
        12
        arrow-down
        5
        ·
        1 month ago

        No we haven’t. We have an appearance of a AI. Large language models and diffusion models are just machine learning. Algorithm statistic engines.

        Nothing thinks, creates, cares, or knows the difference between something correct or wrong.

        • 0laura@lemmy.world
          link
          fedilink
          English
          arrow-up
          13
          arrow-down
          8
          ·
          edit-2
          1 month ago

          I know enough about how LLMs work to gauge how intelligent they are. The reason I have a different opinion than you is not because you or I lack understanding of how LLMs or diffusion models work, its simply that my definition of AI is more “lenient” than yours.

          EDIT: Arguing about which definition is more correct is pointless because it’s totally subjective. However I think that a more lenient definition of AI is more useful in this case, because with more strict definitions we probably never will have something that could be considered AI.

          • Holyginz@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 month ago

            …then we will never have something considered AI then. Making the definition more lenient doesn’t magically make something that isn’t AI into something that is.

            • 0laura@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              3
              ·
              1 month ago

              Or we just use the definition that many people have used for ages and call the code controlling Minecraft creepers are. it’s only recently that everyone has been getting upset about ai being used too much

          • howrar@lemmy.ca
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 month ago

            It’s not completely subjective. Think about it from an information theory perspective. We want a word that maximizes the amount of information conveyed, and there are many situations where you need a word that distinguishes AGI, LLMs, deep learning, reinforcement learning, pathfinding, decision trees and the like from the outputs of other computer science subfields. “AI” has historically been that word, so redefining it without a replacement means we don’t have a word for this thing we want to talk about anymore.

            I refuse to replace a single commonly used word in my vocabulary with a full sentence. If anyone wants to see this changed, then offer an alternative.

          • arglebargle@lemm.ee
            link
            fedilink
            English
            arrow-up
            10
            arrow-down
            9
            ·
            1 month ago

            There is no intelligence. There is only algorithms. The place we are at is not anywhere near approaching artificial intelligence, it is only buzzwords. If you know about how this works this should be clear. I think I was being very objective: we have statistical engines and diffusion formulas. No intelligence, of any kind, is being demonstrated. AI is a marketing term at this point. No original ideas, no real knowledge of past or future events, no ability to determine correct answers from false ones. Even the better models that try to basically watch the other models are still not that great beyond the basics of “what is the next most likely word here”.

            • 0laura@lemmy.world
              link
              fedilink
              English
              arrow-up
              9
              arrow-down
              2
              ·
              1 month ago

              Like I said, that’s where we disagree. I call the code controlling Creepers in Minecraft AI

              • wizardbeard@lemmy.dbzer0.com
                link
                fedilink
                English
                arrow-up
                2
                ·
                1 month ago

                While I think that is a different meaning than the current fad/bubble driven meaning that marketing groups would have people believe AI is, it’s interesting how fast people have forgotten the old uses of the term.

                Personally I try to avoid using it to describe those now, since AI in popular parlance has been extended to at least imply a lot more lately.

        • 0ops@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          1 month ago

          Machine learning is a subset of artificial intelligence, along with things like machine perception, reasoning, and planning. Like I said in a different thread, ai is a really, really broad term. It doesn’t need to actually be Jarvis to be AI. You’re thinking of general ai

      • arglebargle@lemm.ee
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 month ago

        I am fully aware of Alan Turnings work and it is rather exceptional when you read that formulas were be8ng created for diffusion models in the late 40’s.

        But i really don’t care thar whoever wrote that wikipedia page believes the hype. We are still in statistical algorithm stages. Even on the wiki page it says thar AI is aware of its surroundings as a feature of AI. We do not have that.

        Also, it appears that most people are still not fooled by “ai” as we have it today, meaning it does not pass even the most basic Turing test. Which a lot of academic believe is not even enough as a marker of ai ad that too wad from the 50’s

        • 0ops@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          1 month ago

          “Aware of its surroundings” is a pretty general phrase though. You, presumably a human, can only be as aware as far as your senses enable you to be. We (humans) tend to assume that we have complete awareness of our surroundings, but how could we possibly know? If there was something out there we weren’t aware of, well we aren’t aware of it. What we know as our “surroundings” is a construct the brain invents to parse our own “raw sensor data”. To an LLM, it “senses” strings of tokens. That’s its whole environment, it’s all that it can comprehend. From its perspective, there’s nothing else. Basically all I’m saying is that you seem to be taking awareness-of-surroundings to mean awareness-of-surroundings-like-a-human, when it’s much more broad than that. Arguably uselessly broad, granted, but the intent of the phrase is to say that an AI should observe and react flexibly.

          Really all “AI” is just a handwavy term for “the next step in flexible, reactive computing”. Today that happens to look like LLMs and diffusion models.

    • ASeriesOfPoorChoices@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      we don’t have broad/general/strong AI, true.

      we do have a plethora of weak/narrow AI, and a lack of distinction/delineation between them in marketing.