I use ChatGPT just for programming and it gives wrong answers half of the time.
Not picking fights. Just curious.
Is this an improvement or a decline in your overall code programming success?
I am a hobbyist (and not very good) programmer, and while ChatGPT (free version) often gives me wrong answers, it still gives me some insight on how some stuff could be done (intentionally or not) or how something works and is actually somewhat helpful in learning stuff, but I guess this could be double-edged sword even in that regard.
It is also pretty good at detecting simple code errors, from what I have seen.
Overall more positive than negative, but I wouldn’t recommend to use it blindly.
Huge improvement in work flow.
Don’t get it to write your code for you, it’s not gonna work 3/10 times. Instead use it to review your code, help remove any code smells for refactoring.
I don’t use chatGPT, but work with colleagues who do. They’re productivity visibly drops and half the time I gotta fix their shitty code.
I use chatGPT for any topic I’m curious about, and like half the time when i double check the answers it turns out they’re wrong.
For example i asked for a list of phones with screens that don’t use PWM, and when i looked up the specs of the phones it recommended it turned out they all had PWM, even though in the chatGPT answer it explicitly stated that each of these phones don’t use PWM. Why does it straight up lie?!
It’s not lying. It has no concept of context or truth. Auto complete on steroids.
Yeah I also do and it is indeed frequently incorrect. It is good when you have like no idea about what you’re doing. It can help you get on track and then you can research by yourself.
i’m studying mechanical engineering and there’s a guy in our class who’s obsessed with chatgpt. he’s always trying to solve all of the tasks using chatgpt and he’s always the first to share the solution in zoom. so far it’s never been correct but he just sticks with it…
The other half though is sweet times.
During the early newspaper era, they would write the editor with basic questions like they were google.
Yeah I read a 30 year old newspaper a while back and it was like super high latency internet. Message board, posts, replies to posts, personals etc. None of that stuff makes it into newspapers anymore…
Isn’t it still a thing.
It is. I’ve seen ‘Write the Editor’ sections often in the magazines I check out from the library from time to time. IIRC: Popular Mechanics, Popular Science, and The New Yorker have one.
I don’t get why people trust their answers so much. They lie. Confidently. Constantly.
I usually ask the GPT, then look up the topic myself based on terms and keywords that were mentioned
No! You’re meant to blindly trust every source of information you read, AI or not.
What are some other communities that are less used now? WritingPrompts and PhotoshopBattles come to mind for me
Oh man. I forgot that PhotoshopBattles existed.
That’s a good question. I would guess that it’s lessened some. But both are for creative tasks vs explaining a proven topic, item, or thing.
People on ELI5 ask questions that can be answered with a single google search. Yet they do not do the google search. What makes yoi think they will do the bard or chatgpt?
Quota is in shambles
So, no change then?
All they need is a couple bots that rely on GPT for outputs and it’d be like nothing changed.
ELI5 how to use ChatGPT and Bard
Interestingly, as ChatGPT might be trained on these ELI5 questions and as a result they are asked more infrequently, it might get worse over time or out of date on these types of questions by its own doing. I especially wonder how bad this influence will get on subjects that you’d normally search stackoverflow for.