Inspiration
Navigating the sea of internship postings is a massive time-sink—hundreds of roles are scattered across LinkedIn, Indeed, and the web, making it nearly impossible to keep track of the best opportunities. I envisioned a platform that actually streams daily data from across the web to a central hub, making it easy to discover and apply to internships as they go live.
What it does
SpamApply provides a curated, real-time list of internship opportunities across the USA and Canada. The system displays essential details for every role—including job title, company name, location, and the date posted—and provides the direct link of where to apply.
How I built it
Since my primary goal for this hackathon was to master how production-grade pipelines work, I focused heavily on orchestrating the CI/CD workflow. I spent significant time making the process fully automatic using Databricks Jobs and schedules, all managed through Databricks Asset Bundles (DABs) to ensure a professional, scalable deployment.
Challenges I ran into
The engineering journey was not without its hurdles—configuring the Databricks CLI within a local Windows PowerShell environment was a major struggle. I faced repeated difficulties trying to configure the DABs correctly and dealt with high latency problems that threatened to slow down the entire pipeline.
Accomplishments that I'm proud of
I successfully built a fully functional cloud pipeline that can actually run every single day. I managed to take the system from processing messy, raw data to delivering 84-85 high-quality matches specifically tailored for the tech industry.
What I learned
This project taught me the critical skill of bridging the gap between local development and cloud production. I now have a deep understanding of how to take code from a local VS Code environment and turn it into an orchestrated, automated service in the cloud.
What's next for SpamApply
Scale. Optimization. Frontend.
Log in or sign up for Devpost to join the conversation.