Discussion about this post

User's avatar
Luis Pessoa's avatar

Hi Elisa,

Great post! This is a very good list of models.

Personally, I tend to use the 3B/4B models more often on CPU — specifically Gemma3, Llama3.2, and Qwen3. They provide decent responses.

I use Ollama (with Docker) for apps.

I really like LM Studio, but since it has some restrictions at work, I often use Jan for easily testing models in a ChatGPT-like interface. It also offers an OpenAI-like API, which can be handy if needed. Anyway, they are both easy to use.

Expand full comment
1 more comment...

No posts