Model Zoo.

A comprehensive registry of intelligence kernels and cloud APIs officially supported by Luca OS. Local models execute entirely on-device with zero data telemetry. Cloud models require network access and API keys.

Brain / Core Logic

๐Ÿง 
Local

Qwen 2.5

Alibaba โ€ข 7B Parameters
Context32k
Size4.2GB
VRAM6GB
๐Ÿ‹
Local

DeepSeek R1

DeepSeek โ€ข 8B Parameters
Context128k
Size4.7GB
VRAM8GB
๐Ÿฆ™
Local

Llama 3.2

Meta AI โ€ข 3B Parameters
Context8k
Size2.0GB
VRAM4GB
๐Ÿง 
Cloud API

Claude 3.5 Sonnet

Anthropic โ€ข Frontier
Context200k
LatencyLow
ModalityCoding Focus
โšก
Cloud API

GPT-4o

OpenAI โ€ข Frontier
Context128k
LatencyUltra-Low
ModalityMultimodal

Vision & Parsing

๐Ÿ‘๏ธ
Local

UI-TARS Vision

Tencent โ€ข 7B Parameters
RoleGUI Control
CoordinateNative
VRAM8GB
๐Ÿ–ผ๏ธ
Local

SmolVLM

HuggingFace โ€ข 2B Parameters
RoleFast Parsing
Context8k
VRAM3GB

Memory / Vectors

๐Ÿงช
Rank 2

Model2Vec Potion

Local โ€ข 5MB
Size5 MB
Speed500x Faster
RoleMobile Vector
๐Ÿž
Rank 6

MixedBread XS

Local โ€ข 90MB
Size90 MB
QualityBalanced
RoleCore Memory
๐Ÿ—‚๏ธ
Rank 7

BGE Small En

Local โ€ข 130MB
Size130 MB
MTEB62.17
RoleHigh Accuracy

Voice Engine

๐ŸŽ™๏ธ
TTS Rank 8

Kokoro-82M

TTS โ€ข 10 Voices
Size100 MB
RAM Req512 MB
RolePremium Voice
โšก
TTS Rank 1

Piper Amy

TTS โ€ข Ultra-Fast
Size60 MB
RAM Req256 MB
RoleZero-Latency
๐Ÿš€
STT Rank 2

Moonshine Tiny

STT โ€ข Next-Gen
Size190 MB
Speed15x Whisper
RoleFast Listen

Ollama Native Bridge

Luca OS seamlessly interfaces with your existing Ollama installation. If you already have a model running locally (e.g. llama3.2, qwen2.5, or custom fine-tunes), Luca will dynamically detect it and bypass the standard download sequence, booting instantly using your existing hardware-accelerated endpoints.

Auto-Discovery Port 11434
// Startup Initialization
[LUCA] Initiating hardware scan...
[LUCA] Polling existing inference engines
[LUCA] Found endpoint localhost:11434
[LUCA] Discovered 'qwen2.5:latest'
[LUCA] Bypassing native provision. Booting OS...