Abstract created by Good Solutions AI
In abstract:
- PCWorld reviews that Claude AI customers are adopting “caveman” prompting strategies to cut back token consumption by stripping filler phrases and articles from responses.
- This methodology can dramatically reduce output tokens, with one neural community rationalization dropping from 460 to simply 80 tokens utilizing simplified language.
- Whereas efficient for coding duties, debates proceed over whether or not caveman prompting compromises response high quality for nuanced explanations.
It’s no secret that Claude gobbles up tokens like a Corvette guzzles fuel—and similar to fuel, tokens price cash. That’s why the heaviest Claude customers are at all times on the lookout for methods to save lots of on token utilization. The newest trick? A wild and controversial new methodology: talking like a caveman.
Right here’s the way it works: Utilizing both a specialised system immediate or a full-on plug-in, customers are instructing Claude to speak like a caveman, stripping away any and all fluff corresponding to preambles, reward, asides, filler phrases (like “very” and “actually”), redundancies, and even articles (“a,” “the”).
One of many many steered “caveman” system prompts floating round Reddit embrace this one: “Converse primitive. Use nouns and verbs. No grammar filler (the, is, are, of). Preserve phrases quick. Save tokens. Be blunt.”
The outcomes will range wildly relying on the precise immediate you utilize. I attempted a variation of that immediate, including the particular articles to keep away from whereas additionally asking it to cut the pleasantries and filler phrases. What I obtained from Claude wasn’t precisely true caveman communicate, though I’m certain you possibly can get that with correct prompting. As an alternative, I obtained super-clipped responses that had been devoid of idle chit-chat and digressions.
For instance, when asking a caveman-prompted Claude to elucidate the fundamentals of neural networks, it got here again with this:
Neural networks: layers of related nodes. Enter layer takes knowledge. Hidden layers discover patterns. Output layer offers consequence.
Nodes move alerts, weighted by significance. Coaching adjusts weights utilizing backpropagation — decrease error, repeat thousands and thousands of occasions.
Key varieties: CNNs (pictures), RNNs (sequences), Transformers (language). Fashionable LLMs are transformers at huge scale.
That’s a 50-word response, which comes out to roughly 80 tokens. (There are extra tokens than phrases as a result of an AI mannequin should break down bigger phrases, punctuation, and emojis into smaller items earlier than processing them, and every of these items counts as a token.) Evaluate that to the 330-word, 460-token response from Claude after I didn’t use the caveman immediate.
So, case closed, proper? Let’s all use caveman mode! Effectively, not fairly.
Ever for the reason that caveman immediate started gaining traction earlier this month, it has spawned heated debates on Reddit and elsewhere. Some customers argue that this caveman methodology results in dumbed-down responses. Others observe that any token financial savings are just for the output tokens, not the tokens you ship to Claude. Additionally, an elaborate “caveman” system immediate will price you each time you ship a brand new immediate.
It’s additionally doable that the Claude caveman methodology isn’t a jack-of-all-trades. Whereas it might be well-suited for coding, a activity the place you’re sometimes anticipating a cut-and-dried response, it’s in all probability not one of the best match for “explainer” prompts that require extra element or nuance, corresponding to my prior “inform me about neural networks” request.
Nonetheless, the Claude caveman methodology is one other instance of how customers are getting inventive about boosting their AI token effectivity, a difficulty that’s looming for even informal AI customers as increasingly more highly effective agentic instruments (like Claude Cowork) hit the mainstream.
Certainly, the makers of the extra standard Claude caveman plug-ins are shortly discovering methods to stretch their Claude token use with out sacrificing high quality within the discount.

