Benchmarked a local-first MCP code-intel server on gin / nestjs / react — full methodology + reproducer

Posted by Parking-Geologist586@reddit | LocalLLaMA | View on Reddit | 0 comments

I've been working on a local-first code intelligence MCP server and benchmarked it on three pinned public repos. All numbers are reproducible with one command (npm run bench clones the exact versions and runs the profiler).

Repo Files Cold index Search p95 Impact p95 DB size
gin-gonic/gin v1.10.0 99 10s 12ms 0.75ms 4 MB
nestjs/nest v10.4.0 1,709 22s 14ms 0.88ms 11 MB
facebook/react v18.3.1 4,368 152s 26ms 1.18ms 67 MB

Measured on M-series Apple Silicon, no GPU, cold start includes the full index build.

Stack

Why not just a bigger embedding model?

Because the three signals handle different failure modes:

Any one signal on its own has clear failure cases. Fusing them with RRF is scale-invariant and catches the complementary strengths.

Honest weaknesses

What I actually built it for

MCP servers for AI coding agents (Claude Code, Cursor, Windsurf, Google Antigravity) mostly either (a) upload your code to a cloud index or (b) hallucinate file paths because they don't have an actual graph. I wanted something that
gave Claude Code the same mental model of a repo that a senior engineer has — symbol reachability, blast radius, test coverage, structural importance — without anything leaving my laptop.

Technical deep-dive

Install

npm install -g sverklo
cd your-project && sverklo init

sverklo init auto-detects your installed AI coding agent and writes the right MCP config. MIT licensed. Opt-in telemetry (off by default, full schema documented, mirrored to a local log before any network call).

Repo: github.com/sverklo/sverklo

If anyone wants to benchmark sverklo against another local-first tool on the same repos, I'll run whatever comparison you propose and post the numbers in a reply. Interested in what shape of query breaks it most.