Ultra-compressed communication mode for Claude Code. Cuts ~75% of tokens while keeping full technical accuracy.
npx claudepluginhub juliusbrussee/cavemanUltra-compressed communication mode. Cuts ~75% of tokens while keeping full technical accuracy by speaking like a caveman.
A Claude Code skill/plugin and Codex plugin that makes agent talk like caveman — cutting ~75% of tokens while keeping full technical accuracy.
Based on the viral observation that caveman-speak dramatically reduces LLM token usage without losing technical substance. So we made it a one-line install.
</td> <td width="50%">"The reason your React component is re-rendering is likely because you're creating a new object reference on each render cycle. When you pass an inline object as a prop, React's shallow comparison sees it as a different object every time, which triggers a re-render. I'd recommend using useMemo to memoize the object."
</td> </tr> <tr> <td>"New object ref each render. Inline object prop = new ref = re-render. Wrap in
useMemo."
</td> <td>"Sure! I'd be happy to help you with that. The issue you're experiencing is most likely caused by your authentication middleware not properly validating the token expiry. Let me take a look and suggest a fix."
</td> </tr> </table>"Bug in auth middleware. Token expiry check use
<not<=. Fix:"
Same fix. 75% less word. Brain still big.
Sometimes too much caveman. Sometimes not enough:
<table> <tr> <td width="33%"></td> <td width="33%">"Your component re-renders because you create a new object reference each render. Inline object props fail shallow comparison every time. Wrap it in
useMemo."
</td> <td width="33%">"New object ref each render. Inline object prop = new ref = re-render. Wrap in
useMemo."
</td> </tr> </table>"Inline obj prop → new ref → re-render.
useMemo."
Same answer. You pick how many word.
Real token counts from the Claude API (reproduce it yourself):
<!-- BENCHMARK-TABLE-START -->| Task | Normal (tokens) | Caveman (tokens) | Saved |
|---|---|---|---|
| Explain React re-render bug | 1180 | 159 | 87% |
| Fix auth middleware token expiry | 704 | 121 | 83% |
| Set up PostgreSQL connection pool | 2347 | 380 | 84% |
| Explain git rebase vs merge | 702 | 292 | 58% |
| Refactor callback to async/await | 387 | 301 | 22% |
| Architecture: microservices vs monolith | 446 | 310 | 30% |
| Review PR for security issues | 678 | 398 | 41% |
| Docker multi-stage build | 1042 | 290 | 72% |
| Debug PostgreSQL race condition | 1200 | 232 | 81% |
| Implement React error boundary | 3454 | 456 | 87% |
| Average | 1214 | 294 | 65% |
Range: 22%–87% savings across prompts.
<!-- BENCHMARK-TABLE-END -->[!IMPORTANT] Caveman only affects output tokens — thinking/reasoning tokens are untouched. Caveman no make brain smaller. Caveman make mouth smaller. Biggest win is readability and speed, cost savings are a bonus.
A March 2026 paper "Brevity Constraints Reverse Performance Hierarchies in Language Models" found that constraining large models to brief responses improved accuracy by 26 percentage points on certain benchmarks and completely reversed performance hierarchies. Verbose not always better. Sometimes less word = more correct.
npx skills add JuliusBrussee/caveman
Or with Claude Code plugin system:
claude plugin marketplace add JuliusBrussee/caveman
claude plugin install caveman@caveman
Codex:
/pluginsCavemanInstall once. Use in all sessions after that.
One rock. That it.
Trigger with:
/caveman or Codex $cavemanStop with: "stop caveman" or "normal mode"
Sometimes full caveman too much. Sometimes not enough. Now you pick: