Dec 24, 2025
Deploying Qwen 2.5 32B with vLLM Building a dedicated inference server for local AI workloads using vLLM and AWQ quantization
LLM · Infrastructure · Docker · Python
2 posts
Building a dedicated inference server for local AI workloads using vLLM and AWQ quantization
Replacing the Tailscale control plane with a self-hosted alternative for complete network sovereignty
grep: no matches found