Skip to content

Changelog

What's new in Cloaked.

v1.2.0 April 2, 2026

Ministral 3, Tappable Citations & Smarter Memory

Added

  • Ministral 3 3B — new edge-optimized model from Mistral AI with chat, coding, and multilingual capabilities (replaces Mistral 7B — smaller download, runs on more devices)
  • Tappable citation links — inline [1], [2] citations in web search responses now open the source URL directly in Safari
  • Configurable thinking budget — control how long models spend reasoning (256–4096 tokens) with a new Settings slider
  • Continuation generation — when the thinking budget is reached, the model transitions smoothly to producing an answer instead of cutting off
  • Download confirmation — new dialog when downloading models that may run slowly on your device

Improved

  • Smarter model recommendations — Qwen 3.5 2B is now the default for 8GB devices (better speed/quality tradeoff)
  • Leaner model library — removed legacy and redundant models for a focused catalog of 12 models from 5 AI labs
  • Memory management — RAM checks now run before loading models, preventing out-of-memory crashes
  • Onboarding — sub-8GB devices default to Qwen 3.5 0.8B for a smoother first experience
  • Source pills — numbered labels (“1 · example.com”) now map directly to inline citations

Fixed

  • Memory errors now surface specific messages instead of generic “generation failed”
  • Chat template detection for Ministral model IDs
  • Missing translated strings across all 10 non-English languages
  • 4B+ models now correctly show “May run slowly” on 8GB iPhones
v1.1.0 March 14, 2026

Qwen 3.5 Series & Enhanced Reasoning

Added

  • Qwen 3.5 model series — 0.8B, 2B, 4B, and 9B variants with next-generation reasoning
  • Qwen 3 model series — 0.6B, 1.7B, 4B, and 8B with built-in thinking mode
  • Thinking mode toggle — switch between fast responses and chain-of-thought reasoning
  • Expanded model library — now 15 models from 5 AI labs

Improved

  • Model download reliability and resume support
  • Chat response streaming performance
  • Memory management for larger models
  • Model card UI with capability badges

Fixed

  • Occasional crash when switching models mid-conversation
  • Voice input not activating on first tap in some cases
  • Markdown rendering for nested code blocks
v1.0.0 January 15, 2026

Initial Release

Added

  • On-device AI chat powered by Apple’s MLX framework
  • 7 AI models at launch — Llama 3.2, Qwen 2.5, Phi-4 Mini, DeepSeek R1, Mistral 7B
  • Voice input — on-device speech recognition, no cloud processing
  • Text-to-speech output for hands-free use
  • Web search via DuckDuckGo — privacy-preserving, no search history
  • Projects — organize conversations with custom system prompts
  • Persistent memory — context that carries across conversations, stored locally
  • Markdown rendering with syntax-highlighted code blocks
  • 6 accent color themes with light and dark mode
  • Siri Shortcuts integration
  • Full offline support — works in airplane mode after model download
  • Zero accounts — no sign-up, no API keys, no subscriptions