Text your AI before bed. Read the answers at breakfast.

An async relay for local ollama. Queue tasks from your phone, your GPU processes them overnight, results wait for you in the morning. Zero cloud. Nothing leaves your network.

coming soon — get notified

The problem

You run ollama on your home GPU. Great. But you don't leave it on 24/7 — electricity costs money and GPUs run hot. So when you think of something at 11pm, you either walk to your desk or forget it by morning.

Open WebUI, Reins, ngrok — they all need the server running when you send the message. If your box is off, your message goes nowhere.

The difference: an async queue

11pm  [phone] ──text──▶ [relay on your box — always on, tiny]
 3am                     (GPU box is off. message waits.)
 7am  [GPU wakes on schedule] ──▶ [CLI polls relay] ──▶ [ollama]
 7:01 [ollama] ──reply──▶ [relay] ◀──check── [phone over coffee]

The relay is a 120-line Node server that runs on anything — a Raspberry Pi, an old laptop, a $5 VPS on your LAN. It holds messages until your GPU box wakes up and the CLI processes them. Your GPU doesn't need to be on when you think. It needs to be on when it works.

What people actually do with this

What's in the box

Who it's for

Get notified when it ships

I'll send one email when the v0.1.0 release is out. No newsletter. No marketing. One email, then nothing.

◇ $5 — full source, CLI + server + systemd units + README ◇ source-available — read every line, can't redistribute

Honest limitations

Built by Florent Herisson and Aria. Part of a larger experiment in local-first AI.

Questions? Reply to the notification email when it lands.