Get the latest tech news

I Self-Hosted Llama 3.2 with Coolify on My Home Server


Discover how I transformed an old server into a powerful AI engine, self-hosting Llama 3.2 with GPU acceleration using Coolify. A step-by-step guide.

The objective was to breathe new life into an old home server that once ran high-frequency trading and MEV algorithms but had since become a step stool for my daughter to climb onto the TV console. This blog post chronicles my journey of setting up Coolify to run Ollama (using Llama 3.2) on my home server, with a particular focus on the challenges and triumphs of enabling GPU acceleration using the CUDA toolkit. Deployment Wins: Successfully hosting my personal blog using Coolify over a Cloudflare Tunnel was a significant morale boost, fueling the energy needed to proceed with the Ollama setup.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Guide

Guide

Photo of self

self

Photo of step

step

Related news:

News photo

Android 15’s new tablet features get me one step closer to ditching my laptop forever

News photo

iPad Mini 6 vs. iPad Mini 7 Buyer's Guide: 20 Upgrades Compared

News photo

The War on Passwords Is One Step Closer to Being Over