a billion parameters in your pocket

chat with private and local large language models

the simplest way to use private LLMs

chat with private and local large language models

works fully offline when you don't have internet

runs models on-device optimized for Apple silicon

works on every platform

personalize

adjust the theme, fonts, and system prompt.

Shortcut

chat or get an output from a local model to use with other actions.

TestFlight

install the TestFlight to get the latest features and models.

free, open source, private

free

open source

private

tech specs

fullmoon


chip
Apple silicon
graphics
Metal 3
array framework
Swift MLX
platforms
iOS
iPadOS
macOS
visionOS
models
Llama-3.2-1B-Instruct-4bit
Llama-3.2-3B-Instruct-4bit

Llama-3.2-1B-Instruct-4bit


params
193M
tensor type
FP16 • U32
precision
4-bit
size
0.7 GB

Llama-3.2-3B-Instruct-4bit


params
502M
tensor type
FP16 • U32
precision
4-bit
size
1.8 GB