All activity
Eran Sandlerleft a comment
A new version is out with updated models and a PC (with NPU) version too

SelfHostLLMCalculate the GPU memory you need for LLM inference
Eran Sandlerleft a comment
π Hey everyone! Iβm Eran β I built LunaRoute because I got tired of not knowing what my coding assistants (like Claude Code and Codex CLI) were actually doing under the hood. LunaRoute gives you full visibility into your AI coding sessions β every prompt, response, token, and cost - all locally and without slowing anything down. You can even redact or tokenize sensitive data before it leaves...

LunaRouteHigh-perf, secure local proxy for AI coding assistants
LunaRoute is a high-performance secure local proxy for AI coding assistants like Claude Code, Codex, and OpenCode. Get complete visibility into every LLM interaction with zero-overhead passthrough, session recording, and powerful debugging capabilities

LunaRouteHigh-perf, secure local proxy for AI coding assistants
Eran Sandlerleft a comment
Zed is awesome! (and fast). And awesome.
Zed has raised a $32M Series B, led by Sequoia
Chris MessinaJoin the discussion
Eran Sandlerleft a comment
Here is the Mac version: https://selfhostllm.org/mac/

SelfHostLLMCalculate the GPU memory you need for LLM inference
Eran Sandlerleft a comment
Hi all, I'm the creator of SelfHostLLM.org. You can read more about why I created it here: https://www.linkedin.com/posts/e...

SelfHostLLMCalculate the GPU memory you need for LLM inference
Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more. Plan your AI infrastructure efficiently.

SelfHostLLMCalculate the GPU memory you need for LLM inference

