Inspiration

Enterprise IT is broken. Siri can't touch a Linux machine. Alexa sends everything to the cloud. IT teams waste hours on repetitive setup tasks — configuring workstations, pulling repos, managing processes — all things that should take a single sentence. We wanted to build the voice assistant that actually works in a real office.

What it does

Zeph is an AI-powered enterprise voice assistant running on an ASUS Ascent GX10 supercomputer. You speak a natural language command — from your browser, or your iPhone — and Zeph's local LLM plans a full multi-step workflow and dispatches it across your entire network simultaneously.

  • Voice or text input via iPhone PWA or dashboard
  • Local LLM (Ollama + qwen3-coder:30b) plans structured workflows — fully offline, no data leaves the building
  • LAN device discovery — automatically finds every machine on the network
  • Multi-app workspace setup — "set up for dev" opens your editor, terminal, and git client tiled on a new workspace
  • Preset + freestyle workflows — the LLM infers intent, not just pattern-matches
  • Real-time dashboard — live device status, command log, CPU/RAM/GPU stats

How we built it

  • GX10 server: FastAPI orchestrator, Ollama LLM, SQLite device registry, async LAN scanner, React dashboard, iPhone PWA
  • Arch Linux client: Flask agent on each machine, hyprctl for desktop control, whitelist-validated bash execution
  • LLM layer: Structured JSON workflow planning via qwen3-coder:30b running fully locally on the GX10
  • Transport: Pure HTTP over LAN — no SSH, no cloud, no VPN

Challenges we ran into

  • Getting the LLM to reliably output structured JSON workflows for complex multi-step commands required significant prompt engineering
  • OpenDrop (AirDrop on Linux) relies on AWDL and OWL kernel drivers — too unstable for a hackathon demo, scrapped
  • TUI apps like btop and lazygit attach to the current terminal instead of opening in a new window — solved by automatically wrapping them in alacritty
  • Hyprland tiling layout requires careful sequencing and sleep timing between app launches

Accomplishments that we're proud of

  • Fully offline AI agent — the LLM runs entirely on the GX10, zero cloud dependency
  • Natural language to multi-machine workflow execution in under 2 seconds
  • The LLM freestyle reasoning — "hackathon mode" or "focus mode" produces sensible app combos without explicit examples
  • Built solo in 24 hours

What we learned

  • Local LLMs are genuinely capable of structured reasoning when prompted correctly — the gap with cloud models is closing fast
  • Enterprise voice control is a real unsolved problem — nothing on the market does what Zeph does for Linux environments
  • Hardware orchestration at the network level is more accessible than expected with modern async Python tooling

What's next for Zeph

  • Whisper STT on the GX10 — standalone mic input, no phone required
  • Smart light control — GPIO and Govee/Tapo bulb API integration
  • HTTPS + WSS — secure the PWA for production use
  • Multi-machine fleet — broadcast commands to dozens of machines simultaneously
  • OpenCode integration — voice-trigger AI coding sessions on any machine in the network

Built With

Share this project:

Updates