LLM Inference Calculator
calculate requirements and configs for optimal LLM deployment

Project Overview
TypeScript
React
Vite
CSS3
A responsive web application designed to help users determine the hardware requirements for running large language models. It allows users to input model parameters (like context length, batch size, and model size) and select different quantization options to calculate the required VRAM and system memory. The app provides recommendations for GPU configurations and shows whether the model can fit in unified memory or requires discrete GPUs.