Sam Altman says ChatGPT should be ‘much less lazy now’::ChatGPT users previously complained that the chatbot was slacking off and refusing to complete some tasks.
Sam Altman says ChatGPT should be ‘much less lazy now’::ChatGPT users previously complained that the chatbot was slacking off and refusing to complete some tasks.
It’s still weird. That reasoning implies that there is a correlation between promising money and long answers in the training data. Seems plausible at first blush, but where can this be actually seen? It’s hardly ever seen in social media, where similar QA formats exists. It’s certainly not in textbooks, where the real good answers are. OTOH there are a lot of tips promised in completely different contexts.
I’m not saying it’s wrong, but there is definitely a lot of cargo cult in prompting strategies.