Ollama + Apple Silicon = stupid fast. MLX makes local LLM inference hilariously quick on your M-series Mac. https://ollama.com/blog/mlx
Ollama is now powered by MLX on Apple Silicon in preview· Ollama Blog
Today, we're previewing the fastest way to run Ollama on Apple silicon, powered by MLX, Apple's machine learning framework.
ollama