Rhapsody — Project Story
About the Project Rhapsody is a cinematic AI music film exploring rhythm, memory, and emotional synchronization between two digital beings.
The project began as an experiment: Can AI capture the feeling of musical connection, without relying on human performers or traditional filmmaking? What started as a simple test quickly evolved into a full narrative music piece shaped by sound, emotion, and motion. The inspiration came from real human chemistry — the way two people can share a moment without speaking, guided only by breath and rhythm. We wanted to recreate that feeling in a world where the actors don’t exist physically, yet feel emotionally real.
What We Learned Building Rhapsody taught us: AI can generate movement, but emotion must be directed. Rhythm has to be shaped visually — not just heard. Subtlety is everything: micro-expressions, eye flickers, and pacing matter more than spectacle. Consistency is a challenge when characters exist only as data — every frame had to be guided, corrected, and color-controlled to preserve continuity.
How We Built It The process happened in three layers: Performance Layer We first designed the emotional arc: tension → synchronicity → release. This became the backbone of all generation prompts, camera decisions, and timing. Visual Layer Every shot was generated and refined using multi-model pipelines, frame interpolation, and post-correction to maintain facial and lighting consistency. Musical Layer The entire film was cut like a piece of choreography. Beat-matching, silence breaks, motion accents, and transitions were all shaped around the rhythm of the music — making the visuals feel performed instead of produced.
Challenges Faced Keeping the two characters consistent for long sequences Maintaining expressive emotional realism across cuts Aligning visual timing to the musical phrasing Combining multiple AI models without losing continuity Achieving natural “warmth” in faces that don’t physically exist Preventing visual drift, lighting inconsistencies, and over-sharpening artifacts Balancing aesthetic nuance with generative unpredictability
Built With
- color
- elevenlabs
- runway
- timing)-davinci-resolve-(final-grade-+-export)-topaz-/-stable-diffusion-(upscaling
- topaz
Log in or sign up for Devpost to join the conversation.