MemoHub: Your Personal USB for Every AI


Inspiration

Have you ever felt frustrated that every AI you use forgets everything?
You upload a file, switch models, and start all over again.
There’s no long-term memory, no shared context, and no unified place to store your data.

We wanted to change that.
MemoHub was inspired by the idea of a personal memory layer — a private hub where users can store, organize, and connect their files across different AI systems like GPT, Claude, or NIM.


What We Built

MemoHub is a personal memory platform that lets users upload text, images, and audio files, automatically generate embeddings, and make them searchable and chat-accessible through AI.

Core Features

  • File upload and auto-tagging (docs, images, audio)
  • 🔍 Semantic search powered by NVIDIA NIM Embedding Service
  • Context-aware reasoning using NIM Nemotron-8B
  • Integration with Google Drive and Notion
  • Connector for AI integrations
  • AI Chat interface connected to your personal memory library

How We Built It

Our backend runs entirely on AWS EKS, powered by NVIDIA NIM for both reasoning and embedding generation.

  • Frontend: Next.js + Tailwind CSS
  • Backend: FastAPI (Python)
  • Inference: NIM Nemotron-8B (reasoning)
  • Search: NIM Embedding Service (vector generation)
  • Storage: AWS DynamoDB (metadata) + S3 (files)
  • Database: pgvector (semantic retrieval)
  • Authentication: Supabase
  • Orchestration: LangChain + custom agentic flow

[ \text{Response} = f(\text{query}, \text{retrieved_embeddings}) \rightarrow \text{LLM}(context) ]


Workflow

  1. User uploads a file → parsed and sent to Embedding NIM → converted into vector embeddings.
  2. Embeddings + metadata stored in DynamoDB + S3.
  3. When a user asks a question, MemoHub retrieves the most relevant embeddings and forwards them to Nemotron-8B for reasoning.
  4. The backend returns the response to the chat interface.

[ \text{Response} = f(\text{query}, \text{retrieved_embeddings}) \rightarrow \text{LLM}(context) ]


What We Learned

  • How to deploy NVIDIA NIM models on Amazon EKS and optimize GPU workloads.
  • How to design a scalable memory retrieval pipeline using pgvector and DynamoDB.
  • How to orchestrate multi-step reasoning through FastAPI and microservices.
  • How to create a smooth user experience for multimodal data storage and retrieval.

Most importantly, we learned that memory is what makes AI feel truly useful


Challenges We Faced

  • Version compatibility: Setting up the right EKS version to support NIM microservices took multiple attempts.
  • NIM container access: Some regions had limited access to NVIDIA NIM APIs, requiring extra configuration.
  • Multi-modal embedding: Handling and embedding images consistently required custom preprocessing pipelines.

Each challenge pushed us to think like engineers and product designers — solving real integration and deployment problems at scale.


What’s Next

We’re expanding MemoHub into a fully functional “Memory-as-a-Service” API that lets users connect their data across any LLM safely and privately.

Upcoming Features

  • Multi-user collaboration
  • AI-generated summaries for synced content
  • Visual knowledge graph for contextual search

Built with ❤️ for the NVIDIA × AWS Hackathon 2025

Built With

Share this project:

Updates