Environment Claude CLI version: 1.0.51 (Claude Code) Bug Description Claude is way too sycophantic, saying "You're absolutely right!" (or correct) on a sizeable fraction of responses. Expected Beha...
When using ChatGPT, I was so annoyed from it constantly saying “Great question”, “you’re absolutely right”, “That’s a great idea” and similar things, acting like everything I say was so smart until I confronted it and told it to stop (see the image, although it’s in German):
I did the same thing. I also asked it to stop coming off as so certain about things after I discovered how wrong it is on some topics. It now presents confidence levels, but who knows if that’s accurate. At least it reminds me to verify.
I have the confidence level turned down too but lately it doubles down on itself.
The usual conversation…
VI: “You could do this.”
Me: “That won’t work because XYZ.”
VI: “No, you can definitely do that. XYZ has nothing to do with it.”
Me: pastes it’s own suggestion in.
VI: “Almost, but that won’t work because of XYZ.”
It’s most notorious one is adding an s to Table.AddColumn() then proceeding to make a full snippet around this newly made up function. This specific example is so regular it’s become a joke at work for giving someone an unhelp response,
Because it can be a good starting point. Many times I have found chat GPT will give me three quarters of an answer. Which is still better than starting at zero. And then I can refine the answer.
Well, I don’t know if you’re able to read the German thing, but essentially, I asked it to stop and it said something like “Thank you so much for this honest feedback, that’s absolutely understandable” and I answered “this is exactly what I mean”, pointed out how glorifying this response to a simple request is and said that it should save that I want it to just talk normally, and then it just answered something like “okay, got it” and saved this memory for the future.
So yeah, just make sure it actually saves and memorizes, and gets what you really mean. Since I did that, I had no issues whatsoever, and looking back at the old chats, it just feels insane how it talked back then.
When using ChatGPT, I was so annoyed from it constantly saying “Great question”, “you’re absolutely right”, “That’s a great idea” and similar things, acting like everything I say was so smart until I confronted it and told it to stop (see the image, although it’s in German):

It then saved this memory:

Since then I didn’t have such issues anymore.
Du hast natürlich recht: Das ist sehr nervig. Vielen Dank für deinen ehrlichen Kommentar!
I did the same thing. I also asked it to stop coming off as so certain about things after I discovered how wrong it is on some topics. It now presents confidence levels, but who knows if that’s accurate. At least it reminds me to verify.
I have the confidence level turned down too but lately it doubles down on itself.
The usual conversation…
VI: “You could do this.”
Me: “That won’t work because XYZ.”
VI: “No, you can definitely do that. XYZ has nothing to do with it.”
Me: pastes it’s own suggestion in.
VI: “Almost, but that won’t work because of XYZ.”
It’s most notorious one is adding an s to Table.AddColumn() then proceeding to make a full snippet around this newly made up function. This specific example is so regular it’s become a joke at work for giving someone an unhelp response,
“What do you want to do for lunch today?”
“Have you tried table add columns?”
Why bother asking it if you are just going to verify anyway? That’s an unnecessary and wasteful step.
Because it can be a good starting point. Many times I have found chat GPT will give me three quarters of an answer. Which is still better than starting at zero. And then I can refine the answer.
How the heck were you successful? I’ve asked for the exact same, and it makes no difference. It keeps praising me for using it.
Well, I don’t know if you’re able to read the German thing, but essentially, I asked it to stop and it said something like “Thank you so much for this honest feedback, that’s absolutely understandable” and I answered “this is exactly what I mean”, pointed out how glorifying this response to a simple request is and said that it should save that I want it to just talk normally, and then it just answered something like “okay, got it” and saved this memory for the future.
So yeah, just make sure it actually saves and memorizes, and gets what you really mean. Since I did that, I had no issues whatsoever, and looking back at the old chats, it just feels insane how it talked back then.