Get the latest tech news
I Self-Hosted Llama 3.2 with Coolify on My Home Server
Discover how I transformed an old server into a powerful AI engine, self-hosting Llama 3.2 with GPU acceleration using Coolify. A step-by-step guide.
The objective was to breathe new life into an old home server that once ran high-frequency trading and MEV algorithms but had since become a step stool for my daughter to climb onto the TV console. This blog post chronicles my journey of setting up Coolify to run Ollama (using Llama 3.2) on my home server, with a particular focus on the challenges and triumphs of enabling GPU acceleration using the CUDA toolkit. Deployment Wins: Successfully hosting my personal blog using Coolify over a Cloudflare Tunnel was a significant morale boost, fueling the energy needed to proceed with the Ollama setup.
Or read this on Hacker News