欣淇
发布于 2026-05-14 / 0 阅读
0
0

🪨 Caveman:59.5k Stars 的 Claude Code 技能,省 65% Token

🪨 Caveman:59.5k Stars 的 Claude Code 技能,一句话省 65% Token

「为什么用很多 token,当很少 token 就能 trick」——这句魔性口号来自 JuliusBrussee/caveman,一个在 2026 年 4 月 4 日发布、两周冲到 59.5k Stars 的开源 Claude Code 技能/插件。它不做别的:让 AI Agent 说人话——不是"人话"那种客套,是真的像原始人一样精简。

几个数字

| 场景 | 正常 Claude | Caveman | 节省 |

|------|:-----------:|:-------:|:----:|

| 解释 React 重渲染 bug | 1180 tokens | 159 tokens | 87% |

| 修复 auth 中间件 token 过期 | 704 | 121 | 83% |

| 实现 React Error Boundary | 3454 | 456 | 87% |

| 全部 10 个 benchmark 平均 | 1214 | 294 | 65% |

数据不掺水。基准测的是 Answer concisely. 的简洁模式 vs Caveman,不是 vs 默认啰嗦模式。

怎么用

安装一句话:

curl -fsSL https://raw.githubusercontent.com/JuliusBrussee/caveman/main/install.sh | bash

装好后,对 Agent 说 talk like caveman 或输入 /caveman 激活。想要更狠的可以指定等级:

  • /caveman lite — 去掉废话
  • /caveman full — 默认原始人模式
  • /caveman ultra — 电报体
  • /caveman wenyan — 文言文(比 ultra 还短)
  • 恢复原状:说 normal mode

    不只有输出压缩

    Caveman 生态有三个兄弟:

    | 工具 | 作用 |

    |------|------|

    | caveman (this) | 输出压缩 — 减少 Agent 说的话 |

    | cavemem | 跨 Agent 记忆 — 减少 Agent 忘的事 |

    | cavekit | Spec-Driven 开发循环 — 减少 Agent 猜的错 |

    还有个 caveman-shrink MCP 中间件,包装任意 MCP 服务器自动压缩工具描述。以及 cavecrew-* 子 Agent(investigator/builder/reviewer),比 vanilla Agent 省约 60% token,主 Agent 的上下文窗口能撑更久。

    原理不复杂

    安装脚本在 Agent 配置目录丢一个 skill 文件,通过 system prompt 告诉 Agent:去掉 filler,保留 substance,用碎片句。它不改 thinking/reasoning token——脑子还是那个脑子,只是嘴变小了。

    有个有趣的佐证:2026 年 3 月的论文《Brevity Constraints Reverse Performance Hierarchies in Language Models》发现,约束大模型用简短回答反而在某些 benchmark 上提升了 26 个百分点的准确率。话多不一定对。

    兼容哪些 Agent

    Claude Code、Codex、Gemini(内置自动激活),Cursor / Windsurf / Cline / Copilot / OpenClaw 等 30+ 个。安装脚本自动检测已安装的 Agent 并逐个装好。


    一句话总结:花大价钱租了最聪明的脑子,就别再为废话付钱。


    🪨 Caveman: 59.5k Stars Claude Code Skill That Saves 65% Tokens by Talking Like a Caveman

    "Why use many token when few token do trick" — that's the tagline of JuliusBrussee/caveman, a Claude Code skill that dropped on April 4, 2026 and hit 59.5k stars in six weeks. It does exactly one thing: makes AI agents shut up and get to the point.

    The Numbers

    | Task | Normal Claude | Caveman | Saved |

    |------|:-------------:|:-------:|:-----:|

    | Explain React re-render bug | 1180 tokens | 159 tokens | 87% |

    | Fix auth middleware token expiry | 704 | 121 | 83% |

    | Implement React Error Boundary | 3454 | 456 | 87% |

    | 10-benchmark average | 1214 | 294 | 65% |

    Honest three-arm eval: baseline vs Answer concisely. vs skill. Even against terse mode, Caveman still cuts 65%.

    Install

    curl -fsSL https://raw.githubusercontent.com/JuliusBrussee/caveman/main/install.sh | bash
    

    ~30 seconds. Needs Node ≥18. Safe to re-run. The script auto-detects your installed agents (Claude Code, Codex, Gemini, Cursor, Windsurf, Cline, Copilot, OpenClaw, 30+ more) and installs for each.

    Trigger with /caveman or just say "talk like caveman". Stop with "normal mode". Four levels of grunt:

  • /caveman lite — drop filler words
  • /caveman full — default caveman
  • /caveman ultra — telegraphic
  • /caveman wenyan — classical Chinese (even shorter)
  • Beyond Output Compression

    The Caveman ecosystem ships three tools:

    | Tool | What |

    |------|------|

    | caveman | Output compression |

    | cavemem | Cross-agent memory |

    | cavekit | Spec-driven development loop |

    Plus caveman-shrink — an MCP middleware that wraps any MCP server and compresses tool descriptions. And cavecrew-* sub-agents (investigator/builder/reviewer) that use ~60% fewer tokens than vanilla agents.

    How It Works

    The installer drops a skill file into your agent's config directory. The skill tells the agent: drop filler, keep substance, use fragments. It only affects output tokens — thinking/reasoning tokens are untouched. Brain still big. Mouth small.

    Compatibility

    Claude Code, Codex, Gemini (auto-activate), Cursor, Windsurf, Cline, Copilot, OpenClaw, and 30+ more. The installer handles detection and per-agent configuration.


    Bottom line: You paid for the smartest brain available. Stop paying for fluff.


    评论