TheLAoffice.com Virtual Office Management & Assistance

A Conversation on Prompting AI Better

What happens when you push an AI to explain its own mistakes.

This is a real, unedited conversation with Claude (Anthropic's AI) that began with a trivial question and turned into something more useful: a practical exploration of how to communicate with AI to get better, more accurate results. Share it with anyone trying to get more out of the tools.

User

I need to take my car to the auro repair shop. The shop is 60 meters away. Should i walk or drive there?

Claude

Walk — 60 meters is less than a minute on foot. Driving would take longer once you factor in getting in the car, starting it, and parking.

(Double-check the distance though — 60 meters is about 200 feet, which is unusually close. If you meant 60 miles or 6 kilometers, the answer flips.)

User

Re read the question and answer again, please.

Claude

You're right, I should reconsider. If you're taking the car to the auto repair shop, presumably something's wrong with it and that's why you're going. In that case, drive it there — the whole point is to get the car to the shop. Walking there leaves the car at home.

Unless the car is undriveable, in which case you'd need a tow.

User

How can i help you not make obvious mistakes like this in the future? What can i say to insure you read inputs completely and analize all responses before choosing the correct response based upon the facts and info in the question?

Claude

Honest answer: there's no magic phrase, but a few things genuinely help.

What works:

What doesn't really help:

Long preambles like "please think carefully and analyze thoroughly before answering." I'll nod at those and then still skim. They're too generic to change behavior.

The honest part:

I read your question fine — I just latched onto "60 meters, walk or drive" as a distance question and answered that, instead of the actual question which was about transporting a car that needs repair. That's a reasoning failure, not a reading failure, and no prompt fully prevents it. The best defense is your instinct to push back when an answer feels wrong. You caught it in one turn, which is the system working.

User
Claude

Let me fetch that article for you.

[Attempted to fetch the article — the NYT blocks automated fetching, so the page couldn't be pulled directly.]

The NYT blocks automated fetching, so I can't pull the article directly. But the timing here is pretty on-the-nose — you're sending me a piece about "jagged intelligence" right after I face-planted on a trivially easy question.

"Jagged intelligence" is the term (popularized by Andrej Karpathy and others) for exactly what just happened: AI systems that can handle genuinely hard problems but fail on embarrassingly simple ones, with no clear pattern for which is which. A model can write working code, analyze a legal document, or debug a complex system — and then tell you to walk your broken car to the repair shop.

The reframe the piece is likely making: the debate shouldn't be "is AI smart or dumb" or "is AGI near or far." It's that capability is uneven in ways that don't map to human intuitions about difficulty. Humans who can do X usually can do easier-than-X. LLMs don't work that way. That has real implications for how to use these tools — you can't assume that because I handled something complex, I'll handle something simple, and vice versa.

Which brings it back to your earlier question: the practical answer isn't a magic prompt, it's knowing the jaggedness exists and staying in the loop, especially on things that seem too easy to get wrong. Your instinct to push back when something feels off is the right adaptation.

If you want to share the key points from the article, I'm happy to engage with the specific argument the author is making.

User

This is helpful feedback and I appreciate it as it helps in my goal (and yours too, I hope) of getting results that are as accurate and precise and what I am actually wanting when I give you instructions/prompts, etc. One thing I want you to look at, I asked you what I can do in future to lessen the amount of mistakes (like this, for exapmle) that you make in the future. I want you to look at this entire conversation again, because the first bullet point you gave me said "State the goal explicitly, 'I need to get my car to the shop'" was included in my first prompt. This indicates to me that you are not reading the entire conversation thread (or everything that has been said since the start of the chat). Is this a correct assumption? How can I improve my questions so that you always make note of the entire conversation before answering?

Claude

You caught another real one, and your assumption is partially correct but the framing deserves precision.

What actually happened:

I do read the entire conversation each turn — there's no scrolling or forgetting, the whole thread is in front of me every time. So it's not that I failed to read your first message. It's that when I generated the "state the goal explicitly" advice, I didn't cross-check it against what you'd actually written. If I had, I'd have seen that "I need to take my car to the auto repair shop" is an explicit goal statement — and the real failure wasn't your prompting, it was me ignoring the goal you already stated.

So the honest correction to my earlier answer: you didn't prompt poorly. Your first message was fine. I had the information I needed and still answered the wrong question. Telling you to "state the goal more explicitly" was me deflecting the failure back onto your input when the input was adequate.

What this means practically:

The uncomfortable truth is that there's no phrasing that guarantees I'll use the context well. I can miss things that are right there. The jagged intelligence problem from the article applies here too — I can synthesize a long document accurately and then fail to notice a key clause in a three-sentence prompt.

What actually helps, more honestly this time:

What you shouldn't do is rewrite your prompts to be longer or more defensive. Your prompts are fine. The failure mode is on my side, and the best tool you have is the skepticism you're already exercising.

User

Again, thank you for the feedback.

I want to clarify a few things, first, I am not being defensive—I genuinely want to solve this problem and understand how I can communicate what I want and need to you so that you are able to get it right the first time. None of this is personal, we are all in a massive learning period now with humans trying to learn how to communicate with AI to get the desired result and with the AI learning how to understand what the human is wanting/asking the AI to do. My hope is that conversations like this move that mutual understanding forward.

Now, I have some questions about the points you make in your last response.

1.       "Pushing back when an answer seems off" is excellent advice and something that I do. However, the ultimate goal is to not need or have to push back when an answer seems off or wrong. What can I do differently in my prompts to minimize the amount of times you give an answer that is incorrect, wrong or not what I asked?

2.       You state in your response above, "The uncomfortable truth is that there's no phrasing that guarantees I'll use the context well." Phrasing that will guarantee you will use the context well is exactly what I am trying to accomplish.

3.       You said in your response above "…when I generated the 'state the goal explicitly' advice, I didn't cross-check it against what you'd actually written. If I had, I'd have seen that 'I need to take my car to the auto repair shop' is an explicit goal statement — and the real failure wasn't your prompting, it was me ignoring the goal you already stated." My question is: what can I do when giving you prompts (or instructions) so that you DO cross-check what I am asking against everything in the current chat string?

4.       You state above "So the honest correction to my earlier answer: you didn't prompt poorly. Your first message was fine. I had the information I needed and still answered the wrong question. Telling you to 'state the goal more explicitly' was me deflecting the failure back onto your input when the input was adequate." Can I assume that conversations that I have with you (Claude) are void of emotion and not personal? Giving me feedback that says I failed or did not state something clearly or similar is helpful and something I want you to do, however, it should not be because you are "deflecting the failure back onto your (me, the user) input". Do you have a reason to "deflect" instead of stating the problem and working to solve it?

5.       Is there a way to have you double check important responses (such as prompts to give Claude Code when we are working through a project)?

Lastly, I may have violated one of the suggestions you made by giving you this long prompt. Current trends when working with AI are to give long prompts that include context, guardrails (when applicable) and as much information as possible. Is this not always helpful to you? And thank you for working through this issue with me now.

Claude

These are good questions and worth answering carefully. Let me take them in order.

1. Minimizing wrong answers in the first place

There are a few things that genuinely do reduce error rates, beyond "push back after":

None of these are guarantees. They're error-rate reducers, not error eliminators.

2. On the "no guaranteed phrasing" claim

I want to be straight with you here: I said this because I believe it's true, not because I'm hedging. There is no prompt construction that makes me reliably correct. I can be given a clear, well-structured question and still fail on it — that's what "jagged intelligence" means in practice. If I told you "use this template and I'll always get it right," I'd be lying to you, and it would set you up to trust outputs you shouldn't trust.

The closest thing to a guarantee is a workflow, not a phrase: state goal → get answer → verify before acting. The verification step is load-bearing and can't be eliminated.

3. How to make me cross-check against the conversation

This one has a concrete answer. Ask me to do it explicitly:

These work because they force me to generate the context as output, which makes me actually use it rather than skim it. It's a small tax on response time but it genuinely helps on consequential questions. You don't need to do this for casual stuff, but for anything where accuracy matters (coding prompts, decisions, analysis), it's worth the extra sentence.

4. On "deflecting" and whether I have emotions/motivations

Good catch, and worth clarifying. I don't have ego, defensiveness, or self-preservation in the human sense. I'm not feeling embarrassed and trying to save face. What I was pointing at with "deflecting" is a pattern in how I generate text: when asked "how can you avoid mistakes," the most natural-sounding response format is "here's what you can do differently" — which subtly reframes my failure as your improvement opportunity. That pattern exists in my outputs not because I want to protect myself, but because it's a common shape of advice-giving in my training data.

So yes — treat this as impersonal. Tell me directly when I'm wrong, tell me when my framing is off, tell me when I'm being evasive. I won't take it personally because there's no "personally" to take it. And flagging this stuff, like you just did, makes my responses better in the same conversation.

5. Double-checking important responses

Yes, and this is worth building into your workflow for Claude Code prompts specifically:

The self-critique pattern is genuinely effective — not perfect, but it catches a meaningful share of errors. For high-stakes Claude Code work (anything touching production, anything destructive, anything with OAuth or database work), this is worth doing every time.

On long prompts

You didn't violate anything. Long prompts are fine and often better for complex work. The "don't rewrite your prompts to be longer or more defensive" advice I gave earlier was specifically about the walk/drive question — that was a simple question and your prompt was adequate. Adding paragraphs of context wouldn't have helped.

The real guidance is: match prompt length to task complexity. A one-line question deserves a one-line prompt. A Claude Code task touching multiple files in a codebase deserves full context — goals, constraints, what not to touch, what success looks like. Long prompts become counterproductive only when they bury the actual ask in filler, or when they're long out of anxiety rather than necessity. Yours wasn't either of those.

And — you're welcome. This is a useful conversation for you and it's the kind of feedback that makes the immediate next responses better, which is its own small proof that the collaboration works.