SelfHostLLM

SelfHostLLM

Calculate the GPU memory you need for LLM inference

103 followers

Calculate GPU memory requirements and max concurrent requests for self-hosted LLM inference. Support for Llama, Qwen, DeepSeek, Mistral and more. Plan your AI infrastructure efficiently.
SelfHostLLM gallery image
SelfHostLLM gallery image
Free
Launch Team
OS Ninja
OS Ninja
Explore and Learn Open Source using AI
Promoted

What do you think? …

Chris Messina

Built to simplify planning for self-hosted AI deployments.

Unlike other AI infrastructure tools, SelfHostLLM lets you precisely estimate GPU requirements and concurrency for Llama, Qwen, DeepSeek, Mistral, and more using custom config.

B̶u̶t̶ n̶o̶w̶ I̶ w̶̶a̶n̶t̶ t̶o̶ s̶e̶e̶ ̶A̶p̶p̶l̶e̶ s̶i̶l̶i̶c̶o̶n̶ ̶a̶d̶d̶e̶d̶ t̶o̶ t̶h̶e̶ m̶i̶x̶!

Update: Now there's a Mac version too!

Cruise Chen

Love how SelfHostLLM lets you actually estimate GPU needs for different LLMs—no more guessing and overbuying fr. Super smart idea, realy impressed!

Mcval Osborne

Very cool calculator, looking forward to checking this out.

Eran Sandler

Hi all, I'm the creator of SelfHostLLM.org.

You can read more about why I created it here:


https://www.linkedin.com/posts/e...

Eran Sandler

Here is the Mac version: https://selfhostllm.org/mac/

Xiaolong Zhu
@erans is it possible to connect more than one macs?
Eran Sandler

@xiaolong_zhu I know some people like Exo Labs (https://github.com/exo-explore/exo) are doing work to allow you to split bigger models across multiple devices - those can also be multiple macs. Do note that based on how you split it, it may require fast networking like the latest Thunderbolt 5.0 etc - but it certainly can be done.

Joey Judd

No way, this is exactly what I needed! Figuring out GPU memory for LLMs has always been such a headache—super smart to automate it. Any plans to support multi-GPU setups?

Hamletina Tonakanyan

Super useful — sizing GPU memory and concurrency upfront saves a ton of headaches. Love that it works with different models.

12
Next
Last