Foundry

Inspiration

Hardware development today is slow and fragmented. Builders constantly switch between tutorials, guess connections, and debug blindly without real-time feedback. We wanted to eliminate that friction by turning the workspace itself into an intelligent system—something closer to a real-world “Jarvis” that understands what you’re building and guides you as you build it.

What it does

Foundry transforms any physical table into a real-time intelligent engineering workspace.

  • Detects components placed on the surface
  • Builds a live model of the system being assembled
  • Projects step-by-step guidance directly onto the workspace
  • Highlights errors, missing connections, and next actions instantly
  • Adapts in real time as components move or change

It also supports:

  • Hardware learning and education
  • Real-time debugging assistance
  • Interactive system explanation for beginners and engineers

How we built it

Foundry is a real-time perception–reasoning–rendering system.

It runs a continuous loop:

  1. Perception Layer
  • Computer vision detects components and workspace state
  • Hand tracking and spatial mapping maintain live context
  1. Reasoning Layer (AI)
  • GPT-4o (via AIP) interprets intent and generates the next build step
  • Outputs structured JSON diagrams instead of images
  1. Rendering Layer
  • Deterministic SVG-based diagram engine visualizes steps
  • Real-time projection overlays guidance onto the physical workspace
  1. Feedback Loop
  • System updates instantly as components move or are added
  • Re-plans steps dynamically based on current state

Tech Stack

  • AI Reasoning: GPT-4o (via Palantir AIP)
  • Computer Vision: OpenCV + custom detection pipeline
  • Tracking: Hand + object tracking system (real-time state mapping)
  • Backend: Node.js / Python orchestration layer
  • Diagram Engine: Custom SVG renderer (deterministic diagram system)
  • Frontend: Real-time projection UI + interactive dashboard
  • Data Model: JSON-based structured circuit/assembly graphs

System Overview

Layer Function Key Technologies
Perception Layer Detects components + tracks workspace state in real time OpenCV, hand/object tracking
Reasoning Layer Interprets intent and generates next-step structured diagrams GPT-4o via Palantir AIP
State Model Maintains live structured representation of the build JSON graph-based schema
Rendering Layer Converts structured steps into deterministic visuals Custom SVG renderer
Projection Layer Overlays guidance directly onto physical workspace Real-time display system

Challenges we ran into

  • Achieving reliable real-time component detection at small scale
  • Reducing system complexity while maintaining accuracy
  • Avoiding fragile wire-level detection by abstracting into component graphs
  • Camera connectivity and latency optimization across multiple tools and setups

We ultimately shifted from raw visual interpretation to a structured component-state model, which significantly improved stability.

Accomplishments that we're proud of

  • Built a fully interactive physical-to-digital build system
  • Achieved real-time recognition → reasoning → projection loop
  • Created a system that feels intuitive without requiring instructions
  • Demonstrated live adaptive guidance during physical assembly

What we learned

  • Simplicity and structure outperform overly complex vision pipelines
  • Representing systems as graphs is more reliable than pixel-level reasoning
  • Real-time constraints force better system design decisions
  • AI is most powerful when it operates on structured data, not raw outputs

What's next for Foundry

We plan to expand Foundry into a full engineering and education platform:

  • More complex hardware and circuit support
  • Classroom and lab integration
  • Expanded component libraries
  • Deeper simulation + prediction of system behavior
  • Collaboration mode for multi-user builds

Built With

Share this project:

Updates