LLM Inference Calculator
calculate requirements and configs for optimal LLM deployment

portfolio

Project Overview

TypeScript

React

Vite

CSS3

A responsive web application designed to help users determine the hardware requirements for running large language models. It allows users to input model parameters (like context length, batch size, and model size) and select different quantization options to calculate the required VRAM and system memory. The app provides recommendations for GPU configurations and shows whether the model can fit in unified memory or requires discrete GPUs.

Like what you see? Reach out via email to collaborate!

Ranadeep Laskar

© 2025 | All rights reserved.