I’m been running gpt-oss-20b on my Macbook Pro for some time now. I typically run different LLMs on my Macbook using ollama as the basis for this. While I’ve tried various models, this model from OpenAI is the one I found that gave me the best answers to the questions I’ve been giving it. I even prefer the answers I get with gpt-oss-20b on my Mac over those I get from larger SaaS models provided by OpenAI, Anthropic and Google.
I’ve found it very good when it comes to asking it questions concerning lifestyle changes. For instance, I asked it for:
- ideas on how to celebrate a big birthday
- a 12 month plan to do Swedish death cleaning on my house
- a 12 month plan to get my estate in order
- a 12 month plan to get back in shape after a long layoff
I’ve found it not very good at all with questions around food and music:
- coming up with music playlists based on certain criteria (over a third of the songs were hallucinations!)
- coming up with spice mixes for cooking meats (the proportions were way off compared to tested recipes)
Like all LLMs, I found it works well as a means to generate ideas. It works ok or poorly when you start expecting exactitude in the answers. YMMV.
Other things I like about gpt-oss-20b running on ollama on my Macbook:
- I can see it go through it’s reasoning when it comes up with the answer, which is a nice feature.
- For some reason the default output format is one to my liking, and I can easily move it into Microsoft Word or Apple Pages for further formatting
- I especially like it because I can ask personal questions and not have those being captured by a service in the cloud. It provides an extra layer of privacy.
I am running gpt-oss-20b and not gpt-oss-120b due to the capacity of my Macbook Pro. Indeed, even gpt-oss-20b takes up a fair amount of resources on my Macbook, despite it having a GPU to do the heavy lifting. I am not in a rush for these answers though, so it is fine. I’d run 120b if I could, but I have been happy with the response with 20b.
If you have a machine with a GPU and enough RAM and disk space, consider downloading ollama and this model and give it a go. Let me know what you discover!
How generative AI works is likely not how you think it works.
Yep, it’s true. If you have some technical skill, you can download this repo from github: