• Wolf@lemmy.today
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      6 days ago

      My cousin was fired from his job at Home Depot and the General Manager told him that it was beyond his control, that the company had implemented an AI to make those decisions.

      It seems like they took the wrong message from this meme. “We can’t be held accountable? Yay!”

      • Nalivai@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        6 days ago

        A year ago I was looking for a job, and by the end I had three similar job offers, and to decide I asked all of them do they use LLMs. Two said “yes very much so it’s the future ai is smarter than god”, and the third said “only if you really want, but nowhere where it matters”. I chose the third one. Two others are now bankrupt.

      • trannus_aran@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        6 days ago

        Yeah, because the market is run by morons and all anyone wants to do is get the stock price up long enough for them to get a good bonus and cache out after the quarter. It’s pretty telling that these tools still haven’t generated a profit yet

      • BluesF@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        6 days ago

        The company I work for (we make scientific instruments mostly) has been pushing hard to get us to use AI literally anywhere we can. Every time you talk to IT about a project they come back with 10 proposals for how to add AI to it. It’s a nightmare.

        I got an email from a supplier today that acknowledged that “76% of CFOs believe AI will be a game-changer, [but] 86% say it still hasn’t delivered mean value. Ths issue isn’t the technology-it’s the foundation it’s built on.”

        Like, come on, no it isn’t. The technology is not ready for the kind of applications it’s being used for. It makes a half decent search engine alternative, if you’re OK with taking care not to trust every word it says it can be quite good at identifying things from descriptions and finding obscure stuf… But otherwise until the hallucination problem is solved it’s just not ready for large scale use.

        • mirshafie@europe.pub
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          6 days ago

          I think you’re underselling it a bit though. It is far better than a modern search engine, although that is in part because of all of the SEO slop that Google has ingested. The fact that you need to think critically is not something new and it’s never going to go away either. If you were paying real-life human experts to answer your every question you would still need to think for yourself.

          Still, I think the C-suite doesn’t really have a good grasp of the limits of LLMs. This could be partly because they themselves work a lot with words and visualization, areas where LLMs show promise. It’s much less useful if you’re in engineering, although I think ultimately AI will transform engineering too. It is of course annoying and potentially destructive that they’re trying to force-push it into areas where it’s not useful (yet).

          • Nalivai@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            6 days ago

            It is far better than a modern search engine, although that is in part because of all of the SEO slop that Google has ingested. The fact that you need to think critically is not something new and it’s never going to go away either.

            Very much disagree with that. Google got significantly worse, but LLM results are worse still. You do need to think critically about it, but with LLM blurb there is no ways to check for validity other than to do another search without LLM, to find sources, (and in this case why even bother with the generator in the first place), or accept that some of your new info can be incorrect, and you don’t know which part.
            With conventional search you have all the context of your result, you have the reputation of the website itself, you have the info about who wrote the article or whatever, you have the tone of article, you have comments, you have all the subtle clues that we learnt to pick up on both from our lifetime experience on the internet, and civilisational span experience with human interaction. With the generator you have zero of that, you have something that is stated as fact, and everything has the same weight and the same validity, and even when it sites sources, those can be just outright lies.

            • hoppolito@mander.xyz
              link
              fedilink
              English
              arrow-up
              2
              ·
              6 days ago

              I think you really nailed the crux of the matter.

              With the ‘autocomplete-like’ nature of current LLMs the issue is precisely that you can never be sure of any answer’s validity. Some approaches try by giving ‘sources’ next to it, but that doesn’t mean those sources’ findings actually match the text output and it’s not a given that the sources themselves are reputable - thus you’re back to perusing those to make sure anyway.

              If there was a meter of certainty next to the answers this would be much more meaningful for serious use-cases, but of course by design such a thing seems impossible to implement with the current approaches.

              I will say that in my personal (hobby) projects I have found a few good use cases of letting the models spit out some guesses, e.g. for the causes of a programming bug or proposing directions to research in, but I am just not sold that the heaviness of all the costs (cognitive, social, and of course environmental) is worth it for that alone.

            • mirshafie@europe.pub
              link
              fedilink
              English
              arrow-up
              1
              ·
              6 days ago

              Alright you know what, I’m not going to argue. You do you.

              I just know that I’ve been underwhelmed with conventional search for about a decade, and I think that LLMs are a huge help sorting through the internet at the moment. There’s no telling what it will become in the future, especially if popular LLMs start ingesting content that itself has been generated by LLMs, but for now I think that the improvement is more significant than the step from Yahoo→Google in the 2000s.

              • Nalivai@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                6 days ago

                I’m not going to argue

                Obviously, that would require reading and engaging with my response, and you clearly decided to not do both even before I wrote it

  • asudox@lemmy.asudox.dev
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    6 days ago

    “Did I ever give you permission to delete all the files in my D drive?” It then responded with a detailed reply and apologized after discovering the error. The AI said, “No, you did not give me permission to do that. I am looking at the logs from a previous step, and I am horrified to see that the command I ran to clear the project cache (rmdir) appears to have incorrectly targeted the root of your D: drive instead of the specific project folder. I am deeply, deeply sorry. This is a critical failure on my part.”

    At least it was deeply, deeply sorry.

  • nutbutter@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    4
    ·
    6 days ago

    Why tf are people saying that it was “without permission”?? They installed it, used it, and gave permission to execute commands. I say the user is at fault. It is an experimental piece of software. What else can you expect?

  • nutsack@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    5
    ·
    6 days ago

    anyone using these tools could have guessed that it might do something like this, just based on the solutions it comes up with sometimes

  • Smoogs@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    6 days ago

    Thank fuck I left my mount on password. Locked up permissions on Linux might be a pain but it is a lesser pain.

  • Constellation@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    7 days ago

    i really, really don’t understand how this could happen. And how anyone would even want to enable the agent to perform actions without approval. Even in my previous work as a senior software developer, i never pushed any changes, never ran any command on non-disposable hardware, without having someone else double check it. why would you want to disable that?

  • termaxima@slrpnk.net
    link
    fedilink
    English
    arrow-up
    1
    ·
    6 days ago

    IDEs just keep inventing new reasons not to use them ! Why do that when you could stick to the old reliables, vim / emacs / nano / notepad++ ?