For the last year or so, my “prompt library” was a mess of Apple Notes pages and a few snippets in Masscode. Every time I wanted to reuse a prompt with...
The reason is simple: because you can — and even faster now. If you have an Apple Silicon Mac (M1 or later) with 16GB of RAM or more, you can run powerful LLMs...
Unlock 2-3x faster AI on Apple Silicon! This post explores optimizing models with Ollama and MLX, boosting performance for demanding applications.
The reason is simple: because you can. M1 or later Macs with 16GB of RAM or more are powerful enough to run useful models on device without sending any data...