Get the latest tech news
Mlx-community/OLMo-2-0325-32B-Instruct-4bit
OLMo 2 32B [claims to be](https://simonwillison.net/2025/Mar/13/ai2/) "the first fully-open model (all data, code, weights, and details are freely available) to outperform GPT3.5-Turbo and GPT-4o mini". Thanks to the MLX project …
That downloads 17GB to~/.cache/huggingface/hub/models--mlx-community--OLMo-2-0325-32B-Instruct-4bit. To start an interactive chat with OLMo 2: The-o unlimited 1 removes the cap on the number of output tokens - the default for llm-mlx is 1024 which isn't enough to attempt to draw a pelican.
Or read this on Hacker News