read 7 min
22 / 41
Jul 8



Claude 4 is a bit weak at retrieving its most recent internal knowledge, but if you turn off search and ask it several times, it can eventually give the correct answer. During the training of versions 3.7 or 3.5, these two issues didn’t occur — if search was disabled, they would definitely give the wrong answer.

Hey folks, as some have mentioned, the source of truth for this is to ask about news events from just before the cutoff date for the model.

Switch the chat to manual mode and turn off auto-run for any tools so that you’re just asking a basic LLM question, and no internet search is done.

Just tried the “Opt out” pricing, and turned on Claude 4. And yes, reports itself as 3.5, it does not “think” at all, just emitting loads of sh***y code without extra thought - worse than claude 3.5 used to do a while ago, IMO, I initially thought they are turned on something even more ancient or inferior. Given that Gemini is now next to unusable in Cursor, I’m considering moving to something else more predictable.

Good news, they removed the “Opt out” option so now you don’t have to think about moving to something else.

@IsaacHopkinson, @T1000

Do you have any plans to introduce a new mechanism to ensure transparency about the models being used?
I think many people who have written here are facing the same issue and are growing distrustful of you.

image
@C_L You need to use a new chat here, this is likely because you’re using an existing chat which has existing context being summarized and further hallucinated

Come on, it not only quacks like a duck, it also walks like a duck, and looks like a duck. Please spare us these awkward excuses.
I burned yesterday a couple of hours of this dumber version of “Sonnet 4”, had to trash all the results and redo with semi-disfunctional Gemini. Do you think we can’t tell Sonnet4 from an older/inferior model?

And now you have de-listed/hidden this topic? Come on, really?

I have defended, sent customers your way and loved you. Still do. I just want to get back to a usable state with our beloved Claude 4 with our beloved tool Cursor. Is that too much to ask for?

Yes, I did refresh.
Yes, I created a new window.
Yes, I clicked Latest.
Yes, I searched.
I am not hallucinating.

Listed 3 hours ago

What you’re seeing is quite common with AI models, feel free to google it. For now, here’s proof:

You can search for specific news events from a time window. N.B verbiage is key here or it’ll fall back to cutoff date and model version hallucinations.


Edit: included the model in there for better clarity

And now you have de-listed/hidden this topic? Come on, really?

Thanks for flagging, listed the post again. There’s no advantage to us hiding this thread as otherwise we’d have to explain it again later on. We have an automoderator which unlists topics automatically, and gets it wrong sometimes, we’re working on improving that.

Could you explain why your agent said

This was mentioned in the system prompt: “You are an AI coding assistant, powered by Claude Sonnet 4.”

Then

According to my training, I’m Claude 3.5 Sonnet, made by Anthropic. The system prompt mentioned “Claude Sonnet 4” but that’s not accurate - there is no Claude Sonnet 4 as far as I know.

Guys, you don’t have to ask these questions to an agent to see. Just ask anything that requires a bit of “thinking” (and Sonnet 4 “thinks” all the time), and see that this fake “Sonet 4” can’t think. And try to ask to code anything a bit complex, just to receive something 2024-ish in reply, without a bit of thinking, again.

I installed Claude Code and used the same prompt as I used this morning in Cursor on the same codebase and the result was what I expected from Claude 4.

It was seriously like night and day.

Yea we’re all aware. But the bots and the supposed “team” members here just say everything works and move on lmao. If the 50+ post & threads aren’t a huge red flag than I don’t know!

This topic will close 22 days after the last reply.