

The principle that one shot prompts are pretty critical for logic puzzles is well established at this point, has been for well over a year now.
Like I said, this is like someone dragging their lawmower out onto the lawn without starting it, and then proclaiming lawnmowers suck cuz their lawn didnt get cut.
You have to start the thing for it to work, mate, lol.
I get that itd be nice if you didnt have to, but thats not how an LLM works, LLMs are predictive text algorithms which means they need something to start predicting off of as a starting point, thats like their whole shtick.
If you dont give them a solid starting point to work from, you are literally just rolling the dice on if it’ll do what you want or not, because Zero shot prompting is going full “jesus take the wheel” mode on the algorithm.
It’s annoying that marketing and consumers have created this very wrong perception about “what” an LLM is.
When you asks someone “knock knock” and they respond with “who’s there?” thats all an LLM is doing, it’s just predicting what text outta come up statistically next.
If you dont establish a precedent, you’re going full RNGjesus on praying it choose the correct direction
And more important, and I CANNOT stress this enough…
Once an LLM gets the answer wrong, if you keep chasing that thread, it will continue to keep behaving wrong
Because you’ve established the pattern now in that thread that “User B is an idiot”, and told it its wrong, and that means its gonna now keep generating the content of what a wrong/stupid responder would sound like
Consider this thought experiment, if you will:
If I hand a person the incomplete text of a play where 2 characters are talking to each other, A and B, and the entire text is B saying dumb shit and A correcting B, and I ask that person to add some more content to the end of what I’ve got so far, “finish this” so to say, do you think they’re gonna suddenly pivot to B no longer being an idiot?
Or… do you think it’s more likely they’ll keep the pattern going I have established, and continue to make B sound stupid for A to correct them on?
Probably the latter, right?
Thats all an LLM is, so if you already have 3 instances of you telling the LLM “No thats wrong you are dumb”, guess what?
You have literally conditioned it now to get even dumber, so its gonna respond with even more wrong responses, because you’re chasing that thread.
To be honest, I dont really blame Anthropic for this.
The prior ruling against The Internet Archive that made it so they couldnt archive digital copies of books effectively has made it so companies like Anthropic must do it this way.
So I blame the shitty courts for the very stupid ruling.