Inspiration

SpaceVision started from a very practical problem. I am currently doing a home reconstruction and was already using Nano banana pro through the chat UI to experiment with interior design ideas: changing floors, styles, materials, and layouts on images of real rooms.

While Gemini was powerful, the workflow was not. Chat-based interaction made it difficult to analyze specific objects, understand why certain designs failed, and iteratively revise a space in a controlled way.

After seeing the Gemini Hackathon announcement on LinkedIn, it became clear that Gemini already had the core capabilities needed - multimodal vision, reasoning, and image generation - but needed a purpose-built workflow. SpaceVision was created to rebuild the renovation and design process around Gemini instead of treating it as a chat feature.

What it does

SpaceVision is a Gemini-powered spatial design assistant that analyzes real spaces from images, renders, or floor plans and redesigns them intelligently.

Gemini detects objects such as walls, floors, furniture, doors, and windows which users can interact with, then reasons about layout, function, and visual consistency. Users can modify specific elements, apply styles, or provide reference images, and Gemini redesigns the space while respecting existing boundaries.

The system runs a tight loop of analysis → reasoning → revision, allowing Gemini to explain failures, adjust constraints, and improve results over time. SpaceVision also adapts to user preferences and learns from past mistakes through contextual reasoning.

How I've built it

SpaceVision is built around Gemini’s multimodal capabilities and treats the model as infrastructure rather than a feature.

Images and plans are first analyzed by Gemini to extract objects, layout, and spatial structure. This analysis is converted into structured context used during reasoning and generation.

Gemini evaluates generated designs for functional correctness, visual coherence, and alignment with user preferences. When issues are detected, the reasoning layer modifies prompts and generation parameters and produces revised outputs.

The frontend provides a clean UX that allows users to trigger actions directly from analysis or assistant views, making Gemini’s reasoning transparent and controllable.

Challenges I've ran into

One of the main challenges was ensuring that SpaceVision did more than generate visually appealing images. The system needed to understand why a design worked or failed and revise intelligently instead of regenerating randomly.

Another challenge was designing guardrails and UX for moments when Gemini’s output was incorrect. Making failures visible and useful - rather than hidden - required careful prompt and workflow design.

Interpreting abstract inputs like 2D floor plans while maintaining realistic spatial constraints was also a significant challenge.

Accomplishments that I am proud of

Built a full workflow around Gemini’s analysis, reasoning, and image generation capabilities

Created an adaptive system that learns user preferences and improves through feedback

Enabled object-level understanding and boundary-aware redesigns

What I've learned

I've learned that Gemini becomes significantly more powerful when analysis, reasoning, and generation are treated as a continuous loop rather than a single step.

I've also learned that transparency matters - explaining why a design failed builds trust and enables better iteration. Contextual learning from preferences and mistakes can create adaptive behavior without retraining models.

Built With

Share this project:

Updates