Nah, we’ll just devalue everyone’s skill and talent to the point where no one makes enough money. Oh wait.
“just work 8 jobs at the same time lol” this guy, probably
Like the CEO of my company, who’s on the board of directors of other companies and shit. And LLM literally writes the shitty corporate talk scripts for him.
He’s gonna do it instead.
RIP KSP2
True, AI won’t, CEOs will
Yeah, I was gonna say, I would agree with “AI cannot do the job of people” but that’s not what he is saying. He is trying to dismiss the idea that he will inevitably replace workers with AI so that he can avoid backlash while slowly incorporating it into their projects.
Say it louder please.
“Everyone” is hyperbole. “Many people” is correct.
Stochastic plagiarism machine AKA LLM models won’t replace anyone
They already have a 52% percent failure rate and this is with top of the line training data
In my line of work (programming) they absolutely do not have a 52% failure rate by any reasonable definition of the word “failure”. More than 9/10 times they’ll produce code at at least a junior level. It won’t be the best code, sometimes it’ll have trivial mistakes in it, but junior developers do the same thing.
The main issue is confidence, it’s essentially like having a junior developer that is way overconfident for 1/1000th of the cost. This is extremely manageable, and June 2024 is not the end all be all of LLMs. Even if LLMs only got worse, and this is the literal peak, it will still reshape entire industries. Junior developers cannot find a job, and with the massive reduction in junior devs we’ll see a massive reaction in senior devs down the line.
In the short term the same quality work will be done with far, far fewer programmers required. In 10-20 years time if we get literally no progress in the field of LLMs or other model architectures then yeah it’s going to be fucked. If there is advancement to the degree of replacing senior developers, then humans won’t be required anyway, and we’re still fucked (assuming we still live in a capitalist society). In a proper society less work would actually be a positive for humanity, but under capitalism less work is an existential threat to our existence.
This is the exact opposite of my experience. We’ve been using codium in my org and 9/10 times it’s garbage and they will not allow anything that is not on prem. I’m pretty consistently getting recommendations for methods that don’t exist, invalid class names, things that look like the wrong language, etc. To get the recommendations I have to cancel out of auto complete, which is often times much better. It seems like it can make up for someone who doesn’t have good workflows, shortcuts and a premium ide, but otherwise it’s been a waste of time and money.
There is literally a university study that proves over 50% failure in programming tasks. It’s not a rational model, deal with it, get off the Kool aid, and move on.
If you didn’t have an agenda/preconceived idea you wanted proven, you’d understand that a single study has never been used by any credible scientist to say anything is proven, ever.
Only people who don’t understand how data works will say a single study from a single university proves anything, let alone anything about a model trained on billions of parameters across a field as broad as “programming”.
I could feed GPT “programming” tasks that I know it would fail on 100% of the time. I also could feed it “programming” tasks I know it would succeed on 100% of the time. If you think LLMs have nothing to offer programmers, you have no idea how to use them. I’ve been successfully using GPT4T for months now, and it’s been very good. It’s better in static environments where it can be fed compiler errors to fix itself continually (if you ever look at more than a headline about GPT performance you’d know there’s a substantial difference between zero-shot and 3-shot performance).
Bugs exist, but code heavily written by LLMs has not been proven to be any more or less buggy than code heavily written by junior devs. Our internal metrics have them within any reasonable margin of error (senior+GPT recently beating out senior+junior, but it’s been flipping back and forth), and senior+GPT tickets get done much faster. The downside is GPT doesn’t become a senior, where a junior does with years of training, though 2 years ago LLMs were at a 5th grade coding level on average, and going from 5th grade to surpassing college level and matching junior output is a massive feat, even if some luddites like yourself refuse to accept it.
😂
Holy shit you’re right, I’m an idiot. Thanks for helping me shift my perspective.
Take-Two CEO: The idea that AI will make everyone unemployed is ‘the stupidest thing I’ve ever heard’
He doesn’t want everyone unemployed, he just want everyone at Take-Two, except him, to be fired thanks to AI.
If overall humans are unemployed, he got no one to sell.
Close. He wants them employed, but paid less since the job is easier.
Nah, this evil over-paid bastard will have already fired everyone by then. Can we make this prick take the place of Johnny Klebitz, please?
And he is right.
Only in the most strict literal sense is he right. Not EVERYONE will lose their job.
Most people won’t in the foreseeable future. LLMs are in a feedback loop and eat their own outputs, making them effectively worse at what they are doing over time. Also they are a gigantic risk factor.
Wow first time in my life I agreed with a CEO, at least only from reading the headline
Nuance kicks in a little more once you know who this cockroach is.
If by AI he means LLM or generative AI then sure. But LLM and GenAI are not truly AI in the full sense of the meaning. They’re building blocks to it. A mind is more complex. The singularity still approaches at blistering speed.
Edit: downvotes don’t change the fact that these "AI"s are not intelligent. It’s a misnomer by ppl who want to sell you shit. https://bigthink.com/the-future/artificial-general-intelligence-true-ai/
Sounds like you’re mixing up AI with AGI and have no idea of what you’re talking about, like 99% of the people on the internet who suddenly act like they’re data science experts. This article is just taking advantage of the fact that people like you don’t know what “AI” means to get clicks by misdirecting you with improperly worded claims. “True AI” doesn’t mean anything.
Also the term “AI” to describe complex algorithms existed long before the technology was ever in the capitalist market. You literally just completely made that part up. One of the guys that coined it (John McCarthy) was one of the most important computer scientists of all time, who was also a cognitive scientist, he’s the same guy who invented garbage collection and Lisp. One of the other guys to coin the term was Claude Shannon, who is widely considered the father of information theory and laid the foundation for the Information Age. The other people to participate in coining the term include the person who made the first assembler & designed the first mass-produced computer, and the guy who proposed the theory of bounded rationality. The guys who coined AI and founded/established the field were pretty much Turing’s successors, not people looking to “sell you shit”.
You mean they aren’t General Artificial Intelligence. They definitely are AI.
That said, I do think we’re like 2 or 3 paradigm shifts away from General AI.
I think we’re probably a lot further than that.
Tbf it probably depends what you would say counts as a “paradigm shift”, but either way, yours is for sure a fair opinion.
LLM are parlour tricks that impress the gullible and easily confused. LLM aren’t building blocks for anything but energy consumption and greed. “The singularity” isn’t coming, true intelligence isn’t coming. You’ve been lied to, and this is the fourth time in my career alone that this has had to be said.