A lightweight, automated solution to host Llama 3.2 (3B) and BitNet models on Proxmox LXC containers. Optimized for CPU inference on low-resource hardware. Includes OpenAI-compatible API setup.