MCP server for LLM quantization. Compress any HuggingFace model to GGUF, GPTQ, or AWQ format. 6 tools: info, check, recommend, quantize, evaluate, push. Self-contained Python server — no external CLI needed.
Install from
M8ven verifies MCPs across every public registry — install directly from whichever one you prefer.
[](https://m8ven.ai/mcp/shipitandpray-mcp-turboquant-0f1790)