• Buelldozer@lemmy.today
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    2 hours ago

    Since we’re whipping out credentials, I’ve been in IT almost 30 years and I can tell you it’s not going to work like that.

    I’m not the person you were replying to but I’ve also been in tech since 1996 and lots of things have worked just like that. All successful technology starts off barely functional and improves over time until nearly all members of it’s intended audience can successfully use it.

    As an example in 1996 setting up a router was a specialty task that required training, by 2016 any moron could buy one off the shelf and have it running in an hour. As another example basic HTML was a specialty skill in 1996 but by 2003 you could do it with Microsoft Word. Smartphones are another example, they went from barely functional Windows Mobile and Blackberry devices which required ridiculous amounts of back end skill to deliver email to iPhones and Androids that any numskull can use for nearly anything at all.

    My point is this; too many people are stuck on the “What use is a newborn baby?” question without realizing that the infant is growing-up at blinding speed. It’s also the first technology to carry the promise, real or not, of self-improvement when it reaches sufficient maturity. Assuming that happens all further improvement will be increasingly automatic and happen even faster.

    AI isn’t going away and it’s only going to get better as time goes on.

    • Shizzymcjizzles@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 hour ago

      You get it. I don’t understand the people in tech burying their heads in the sand. If the question were AGI that is definitely disputable in terms of even the viability. But plain old AI is already here. It’s not even a baby anymore.