Inspiration
LifeReel started from a simple frustration: life can feel deeply vivid in the moment, but the things we use to “capture” it are flat. A quick photo shows what happened, not what it felt like. A voice note preserves words, but not the atmosphere. We wanted to bridge that gap between real experience and artistic expression by giving everyday people a way to turn fragmented memories into something that feels as cinematic as it was emotionally—without needing writing, editing, or filmmaking skills. Our north star was: Everyone has stories. Every story deserves to be told.
What it does
LifeReel is a “cinematic journal” that generates an 8 second trailer-style video from a slice of someone’s life story. Users can speak or type a reflection, optionally add a photo as a visual anchor, and select an “Atmosphere” (for example, Nostalgia, Triumph, or Melancholy). LifeReel then transforms that raw input into a short film sequence with a clear narrative arc (setup → moment → reflection), cinematic composition, and mood-matched visuals. It is designed for private reflection by default, with sharing as an option.
How we built it
We built LifeReel as a lightweight pipeline that turns messy, real-world diary input into production-ready video prompts:
- Gemini 3 Flash as the “Cinematic Intelligence Layer”: It distills voice or text into the emotional core, extracts concrete details, and produces a structured “Director’s JSON” (shot plan, lighting cues, lens language, pacing, and style instructions).
- Identity and context anchoring (multimodal): A reference photo and long-context reasoning help keep the protagonist and details consistent across shots, with a fallback to POV-style shots if likeness generation fails.
- Veo for video generation: The Director’s JSON is translated into high-fidelity prompts for Veo to render the final cinematic frames.
- Frontend: React / Next.js for an immersive, minimal interface focused on the “Live Catch” flow (record → review transcript → snap a visual anchor → choose atmosphere → generate). ## Challenges we ran into
- Emotional fidelity: The hardest part was not generating “a nice video,” but generating a video that matches the vibe of what the user meant, including subtext.
- Avoiding aesthetic sameness: Cinematic styles can easily collapse into repetitive “template” outputs. We tackled this by grounding prompts in specific nouns and details from the user’s real input, not generic mood labels.
- Privacy and trust: A tool that works with diary-like content must be privacy-first and transparent about what is processed and why.
- Identity consistency and failure handling: When character likeness breaks, it can ruin immersion. We designed fallback behaviors (for example, first-person perspective shots) to preserve continuity.
Accomplishments that we're proud of
- We defined a clear product thesis: cinematic memory augmentation that preserves the invisible emotional layer of life, not just the visuals.
- We designed a concrete, end-to-end user flow (“Live Catch”) that feels natural: speak honestly first, then let the system direct the film craft.
- We created a strong technical concept of a Director’s JSON as the bridge between human emotion and video generation, making the system more controllable and scalable than ad-hoc prompting.
- We addressed real risks early (privacy, memory distortion, homogenized aesthetics) instead of treating them as “later problems.” ## What we learned
- Multimodal AI is most powerful when it has a role: framing Gemini as a director helped us design better interactions and more coherent outputs.
- Emotional experience needs structure to translate well into visuals. A three-shot storyboard and a simple narrative arc made results feel intentional instead of random.
- Trust is a feature: for personal content, it is not enough to be impressive. The product has to be transparent, respectful, and clearly labeled as “artistic enhancement.”
- “Cinematic” is not one style. Without careful grounding in user-specific details, cinematic generation quickly becomes generic. ## What's next for LifeReel
- Build a working prototype of the full pipeline (voice/text → Director’s JSON → Veo render) with a tight iteration loop on emotional accuracy.
- Expand “Atmospheres” into a richer set of controllable creative levers (pacing, camera language, color palette) while keeping the UI simple.
- Add optional recap experiences like Season Finale using stored metadata to keep costs manageable.
- Strengthen privacy UX: clearer transparency controls, on-device options where possible, and explicit labeling to prevent confusion between factual memory and artistic interpretation.
Built With
- css
- gemini3
- google-cloud
- googleaistudio
- html
- react
- typescript
- veo3
Log in or sign up for Devpost to join the conversation.