ChatGPT behaves very much like the AI we’ve seen in fiction. You can use it pretty much exactly like the crew of the Starship Enterprise uses the ship’s computer.
In personally trying to use ChatGPT 4 for a job task (programming), I would disagree strongly with this sentiment. I have yet to find a task where it doesn’t partially fail due to no notion of the concepts underlying the topic.
In an example, I asked it to write an implementation of reading from a well known file type as a class. It had many correct ideas for certain operations (compiled from other sources of course), but failed with the basic concept of class instantiation. It was calling class methods in the constructor, which is just not allowed in the language being used. I went through several iterations with it to avail before just giving up on it.
In “normal” language tasks, it seems to be quirky, but passable. But if you give it a highly technical task where nuance and conceptual knowledge are needed? I have yet to see that work in any reliable capacity.
I use it for programming a lot too. You have to explain everything to it like you would a brand new engineer, and then it is often wrong with certain parts like you said. But if you know enough about coding to figure out where it’s wrong, and just write those parts yourself, it can still be a huge time saver.
Yeah, I’d agree that with sufficient iterations and clarifying remarks ChatGPT can produce something close to functional. I was mostly disagreeing with the original comment’s sentiment that it could be treated like the computer on the Enterprise. While they had several plot specific flaws, the duotronic computers were generally competent and didn’t need everything spelled out for them.
ChatGPT behaves very much like the AI we’ve seen in fiction. You can use it pretty much exactly like the crew of the Starship Enterprise uses the ship’s computer.
Edit: star ship to starship
In personally trying to use ChatGPT 4 for a job task (programming), I would disagree strongly with this sentiment. I have yet to find a task where it doesn’t partially fail due to no notion of the concepts underlying the topic.
In an example, I asked it to write an implementation of reading from a well known file type as a class. It had many correct ideas for certain operations (compiled from other sources of course), but failed with the basic concept of class instantiation. It was calling class methods in the constructor, which is just not allowed in the language being used. I went through several iterations with it to avail before just giving up on it.
In “normal” language tasks, it seems to be quirky, but passable. But if you give it a highly technical task where nuance and conceptual knowledge are needed? I have yet to see that work in any reliable capacity.
I use it for programming a lot too. You have to explain everything to it like you would a brand new engineer, and then it is often wrong with certain parts like you said. But if you know enough about coding to figure out where it’s wrong, and just write those parts yourself, it can still be a huge time saver.
Yeah, I’d agree that with sufficient iterations and clarifying remarks ChatGPT can produce something close to functional. I was mostly disagreeing with the original comment’s sentiment that it could be treated like the computer on the Enterprise. While they had several plot specific flaws, the duotronic computers were generally competent and didn’t need everything spelled out for them.