MLX Models

Qwen 3.5, GPT OSS, Ministral — run open-source LLMs locally on Apple Silicon.

Apple MLX lets you run powerful open-source LLMs directly on your Mac for AI text processing — completely free, fully private, no internet required. Models are optimized for Apple Silicon with Metal GPU acceleration and managed directly inside VivaDicta with one-click download.

Qwen 3.5 — Latest Models

VivaDicta includes the newest Qwen 3.5 model family from Alibaba — one of the strongest open-source LLM families available. Qwen 3.5 delivers excellent quality for text processing, translation, summarization, and coding tasks.

ModelBest ForMin RAM
Qwen 3.5 4BFast processing on any Apple Silicon Mac16 GB
Qwen 3.5 9BBest balance of speed and quality32 GB
Qwen 3.5 27BHigh-quality results, near cloud-level48 GB
Qwen 3.5 35B-A3B (MoE)Large model with fast inference via Mixture of Experts48 GB

Other Available Models

ModelBest ForMin RAM
OpenAI GPT OSS 20BOpenAI's open-source model, strong general performance32 GB
Ministral 3BLightweight, fast on 16 GB Macs16 GB
Ministral 8BMistral's compact model, good quality32 GB

Quantization Variants

Most models offer three quantization levels — choose based on your available RAM and quality needs:

  • Base (4-bit) — smallest download, fastest inference, slightly lower quality. Best if RAM is tight.
  • Med (6-bit) — balanced quality and speed. Recommended for most users.
  • High (8-bit) — best quality, closest to the original model. Requires more RAM.

Getting Started

  1. Open VivaDicta settings → AI Provider Apple MLX.
  2. Browse models organized by your Mac's RAM tier (16 GB / 32 GB / 48+ GB). A Recommended badge highlights the best model for your hardware.
  3. Click Download — models are downloaded once and stored locally.
  4. Select the downloaded model as your active AI provider.
  5. Done — all AI processing now runs locally on your Mac.

Managing Models

  • The model browser shows your Mac's memory, storage used by models, available disk space, and download count.
  • Delete models you no longer need to free up disk space.
  • Switch between downloaded models anytime — no re-download needed.

System Requirements

  • Apple Silicon Mac (M1 or later) — required for MLX.
  • RAM — varies by model (see tables above). The model browser shows which models fit your Mac.
  • Disk space — models range from ~2 GB to ~20 GB depending on size and quantization.

MLX vs Cloud AI

FeatureMLX (Local)Cloud (Claude, GPT, etc.)
CostFree foreverAPI fees or subscription
Privacy100% on-deviceText sent to provider
InternetNot needed (after download)Required
QualityGood to excellent (model-dependent)Excellent
SpeedDepends on Mac hardwareFast (server-side)

For a comparison with Apple Intelligence and Ollama, see Local AI Processing.