HomeBlogAboutTools

Introducing LLM Comparator

AI

I built a small tool that I keep wishing existed whenever Iโ€™m comparing LLMs

๐—Ÿ๐—Ÿ๐—  ๐—–๐—ผ๐—บ๐—ฝ๐—ฎ๐—ฟ๐—ฎ๐˜๐—ผ๐—ฟ

LLM Comparator Screenshot

๐—ช๐—ต๐˜† ๐—ถ๐˜โ€™๐˜€ ๐˜‚๐˜€๐—ฒ๐—ณ๐˜‚๐—น

Comparing models is harder than it should be. You end up running the same prompt across different tools, losing history, and struggling to make a clean side-by-side assessment of output quality, latency, and token usage.

This tool lets you run the same prompt set across multiple models and compare results in one place.

Importantly it lets you take a conversation from one LLM and continue it in another LLM. This is useful in cases where you want to use a powerful LLM for the initial prompt but donโ€™t necessarily need it for subsequent refinements.

๐—ž๐—ฒ๐˜† ๐—ณ๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ๐˜€

  • Side-by-side comparisons across multiple LLMs/providers
  • Tracks latency + token usage
  • Save/load sessions as JSON
  • Share sessions via GitHub Gists

๐—ฃ๐—ฟ๐—ผ๐˜ƒ๐—ถ๐—ฑ๐—ฒ๐—ฟ๐˜€ ๐˜€๐˜‚๐—ฝ๐—ฝ๐—ผ๐—ฟ๐˜๐—ฒ๐—ฑ

  • OpenAI
  • Anthropic Claude
  • Google Gemini
  • OpenRouter
  • Local Models (Ollama or any compatible endpoint)

๐—ฃ๐—ฟ๐—ถ๐˜ƒ๐—ฎ๐—ฐ๐˜† ๐—ฏ๐˜† ๐—ฑ๐—ฒ๐—ณ๐—ฎ๐˜‚๐—น๐˜

Thereโ€™s no backend storing your prompts or API keys. Everything is stored locally in your browser (using local storage), and calls are made directly from your browser to the model provider.

๐—ฆ๐—ต๐—ฎ๐—ฟ๐—ถ๐—ป๐—ด ๐˜€๐—ฒ๐˜€๐˜€๐—ถ๐—ผ๐—ป๐˜€

Your most recent session is saved locally in your browser, and you can export it as a JSON file (and re-import it later to pick up where you left off).

If you want to share a comparison, you can also publish the session as a GitHub Gist (public or secret) using a GitHub token, then send a link that will load the exact same session for someone else.

Hereโ€™s an example based on the Claude Career Coach prompt.

๐—ข๐—ฝ๐—ฒ๐—ป ๐—ฆ๐—ผ๐˜‚๐—ฟ๐—ฐ๐—ฒ

This is all a single HTML page with Javascript embedded in it. Download it and deploy it where ever you want, or use it my version.

Source code: Github