ForgeAI
Making AI models accessible for consumer hardware.
We quantize large AI models, fix compatibility issues, and document everything so the community can benefit.
What We Do
- Quantize large models to run on GPUs with 16 GB VRAM or less
- Fix compatibility issues with modern PyTorch, transformers and torchtune versions
- Provide ComfyUI integration for creative AI workflows
- Document all fixes with copy-paste code examples
Our Models
| Model |
What it does |
VRAM |
Link |
| HeartMuLa-3B-4bit |
AI music generation from lyrics |
16 GB |
Download |
Tech Stack
- PyTorch + CUDA
- bitsandbytes (NF4 / INT8 quantization)
- Hugging Face Transformers
- ComfyUI custom nodes
Get In Touch
Open an issue or discussion on any of our model repos.