💡 Inspiration
Every AI storyboard tool in 2026 generates frames. Pretty pictures.
But nobody is solving the actual bottleneck — direction.
Creative teams sketch a board, then scatter the real work across Google Docs, Slack threads, and verbal feedback that evaporates between review cycles. By the time production starts:
- 🔴 Continuity is broken
- 🔴 Pacing intent is lost
- 🔴 The team is polishing the wrong beat
We built the tool that should have existed years ago — a storyboard workspace where direction, continuity, and revision stay attached to the board, not buried in a separate brief.
🎬 What it does
SketchMotion is a collaborative storyboard workspace with a GLM 5.1-powered Director Workflow.
🖼️ The Canvas
Arrange storyboard frames, set durations, add motion notes, collaborate in real time with live cursors and presence.
🎯 Direction Studio
Save creative intent alongside the board — mood, pacing, camera language, lighting, color, continuity constraints, and an avoid-list. This replaces the scattered direction notes that normally get lost between reviews.
🧠 GLM Director Workflow
One click sends the full board context to GLM 5.1. The model reasons over everything in a single structured pass and returns:
| Output | What it does |
|---|---|
| 📊 Storyboard Analysis | Frame-by-frame narrative breakdown |
| 🎥 Shot Plan | Camera direction + timing recommendations |
| 🔗 Continuity Rules | Catches drift before the team wastes time |
| 🚀 Render Strategy | Production handoff guidance |
♻️ Revision-Aware Planning
Type one concise note → hit Apply Revision → GLM 5.1 updates the entire plan without resetting from zero.
The board stays. The direction sharpens.
🛠️ How we built it
🖥️ Frontend React + TypeScript + Tailwind CSS + Vite — infinite canvas with pan/zoom, frame CRUD, sketch editing, drag-and-drop ordering, and real-time collaboration via Supabase Realtime.
⚡ Backend Supabase handles auth, Postgres persistence, file storage, and edge functions. Three edge functions power the AI pipeline:
storyboard-plan— the Director Workflowgenerate-video— video generation pathcheck-video-status— async status polling
🧠 GLM 5.1 Integration
The Director Workflow runs server-side through the storyboard-plan edge function → calls GLM 5.1 via the Z.ai coding endpoint (/api/coding/paas/v4) with thinking enabled → returns a complete director's plan in one pass. Secrets never touch the client.
🏗️ Architecture Decision
Feature-flagged the GLM path (VITE_AI_PROVIDER=zai) alongside the existing Google/Gemini workflow. The core product stays stable. The GLM path is metadata-first — no image fetches, no multimodal credits. Reliable and fast to demo.
🌐 Deployment Vercel (frontend) + Supabase hosted (edge functions + database)
🎯 Why GLM 5.1 specifically
This project demands long-horizon structured reasoning. The model needs to hold:
- 🔄 Frame-to-frame relationships across the full sequence
- 🎨 Creative constraints from 6 different director controls
- ⏱️ Pacing logic that spans every beat
- 📝 Accumulated revision context from prior passes
…all in a single pass.
Prompt chaining would lose coherence. Multi-call orchestration would fragment the plan. GLM 5.1's coding endpoint with thinking enabled is what makes revision-aware planning possible without resetting context.
⚔️ Challenges we ran into
🔧 Provider abstraction — Supporting both Google/Gemini and Z.ai/GLM required a clean provider interface so the UI, edge functions, and state management don't care which model is active. Took several iterations to get right.
📐 Metadata-first pivot — Early versions tried fetching frame images for multimodal analysis. Slow, expensive, fragile. We pivoted to metadata-first (frame titles, order, durations, motion notes) — which turned out to produce better structured plans because it forces the model to reason about sequence and intent, not pixel content.
🔁 Revision coherence — The revision loop needed to feel like updating a document, not starting over. Solved by passing the full previous plan as context alongside the revision note, letting GLM 5.1 decide what to preserve and what to update.
⏰ Time pressure — Built in one week alongside coursework. Prioritizing one tight demo path — board → direction → plan → revision — kept scope sharp.
📚 What we learned
✅ GLM 5.1's long-context reasoning is genuinely strong for structured creative planning — you can't fake this with prompt chaining
✅ Metadata-first AI workflows are more reliable and faster than multimodal approaches for planning tasks
✅ Feature-flagging AI providers early saves enormous debugging time
✅ The best hackathon demos show one tight loop done well, not ten features done halfway
🔮 What's next for SketchMotion
| Priority | Feature |
|---|---|
| 🎬 | Connect GLM shot plan output → automated video generation for a full storyboard-to-animatic pipeline |
| 🖼️ | Multi-frame render orchestration using the render strategy GLM already produces |
| 👥 | Team review workflows where reviewers annotate the plan directly |
| 📄 | Export the director's plan as a production-ready PDF handoff document |
SketchMotion — Direct the storyboard before the render.
Log in or sign up for Devpost to join the conversation.