If that was a jab it my writing then yes, I am absolutely being sincere because I am an expert on this topic. LLMs went from being ok at one-shoting a function a to being so good at hacking that it's difficult to evaluate them. Prospective customers get back to us after a demo and tell us about the exploits it found on their services that are so vague and technical that they wouldn't think to look for them.
> Prospective customers get back to us after a demo and tell us about the exploits it found on their services that are so vague and technical that they wouldn't think to look for them.
Um, have you actually verified that those are actual exploits then? Vague and technical sounds exactly like a description of AI slop...
Just wait until you see the same showing up in compliance realms...
Edit: to be slightly less implicit, consider the cargo cult madness that erupts from people thinking they can address risk management and compliance by auto-generating documentation and avoid really doing the legwork.
I just don't think that's true anymore. Netflix isn't going anywhere. Twitter and Reddit have taken highly visible user-hostile actions since back in 2023 and people stayed. People have become too passive and docile to switch anymore. The portion of the population that's discerning enough to leave is small.
> Groups of people can produce more than lone wolves.
It's not a linear scale. A lone wolf can't produce the latest Assassin's Creed game. A committee can't produce Stardew Valley or Balatro. They're different capabilities, not a simple matter of more/less.
How could anyone who understands this technology call it "artificial intelligence" and create products using it specifically marketed for legal and medical work?
Some people are lazy but properly embarrassed after they're caught, others are idiots, still others have no conscience with which to be bothered.
But it can't be ignored that AI companies shout to the world in their massive marketing campaigns that they'll save you so much time and effort, and then whisper much later we might get some things wrong, check the output. But it's obvious they want everyone to think it works well now.
reply