๐ŸŽ‰ Inspiration

We were struck by the lightning bolt โšก of inspiration when we discovered the incredible potential of Retrieval-Augmented Generation (RAG) and the mind-blowing scalability of Databricks. We thought, why not combine these two superpowers to create a RAG-based LLM model that can swoop in like a superhero ๐Ÿฆธโ€โ™‚๏ธ and save the day for Husky's field service engineers?

Company Website - Husky Webpage

About the Company - Husky Technologies specializes in Injection Molding systems and excels in providing top products and services for industries like consumer goods, medical, beverages, and automotive. They focus on delivering high-performance, efficient solutions globally, with extensive support including installation, training, and maintenance.

Problem Statement - To enhance productivity for Husky Technologies' field service agents, implementing a Large Language Model (LLM) or a context-aware Q&A bot could be transformative. This technology would allow agents to quickly navigate complex documentation necessary for diagnosing equipment issues, significantly speeding up the process. The main challenge is integrating Husky's large knowledge database and internal knowledge into the LLM to ensure it can effectively retrieve and interpret specific information.

๐Ÿค– What it does

Our RAG-based LLM model is like a genius librarian ๐Ÿ“š that lives inside Databricks. It has memorized Husky's entire collection of technical documents, User manuals, Service manuals and can provide spot-on answers to any question thrown its way. It's like having a pocket-sized expert ๐Ÿง  that engineers can consult anytime, anywhere!

๐Ÿง… How we built it

  • Resources were created on Azure cloud including Resource groups, Databricks workspace, Unity Access Connector, Data lake storage (Metastore storage), Azure Key Vault Access (Secret Storage), cluster configurations were setup initially
  • Connections and access management between Azure storage container and databricks workspace were established
  1. Ingested the Pdfs as into Unity Catalog volume store as raw data
  2. Split the pdfs in small chunks of text
  3. Computed the embeddings using a Databricks Foundation model - (bge-large-en) as part of our Delta Live Tables
  4. Created a Vector Search index based on our Delta live Table
  5. Trained the model with input and output examples
  6. Registered the fine tuned LLM model
  7. Endpoint was created and Served using ML-flow endpoint creation
  8. Established a space on hugging face for Chatbot User Interface with Gradio

๐Ÿฝ๏ธ Dataset Used:

Documents used for RAG memory:

Hot Runner Product Handbook [238pgs]

Ultrasync Service Manual [78pgs]

Ultrashot Service Manual [186pgs]

Training Course Doc [20pgs]

๐Ÿ˜… Challenges we ran into

Ensuring that the model's responses made sense and stayed on topic was like herding cats ๐Ÿฑ. Husky's documents are complex, and getting the retrieval process and model architecture just right took more trial and error than a mad scientist's lab ๐Ÿงช. But we persevered, fueled by coffee โ˜• and determination! And also obviously limited Budget ๐Ÿ’ธ

๐Ÿš€ What's next for our RAG-based LLM on Databricks

The sky's the limit! ๐ŸŒŸ

  • โœ… We plan to keep expanding the knowledge base, like a sponge soaking up water ๐Ÿงฝ.
  • โœ… Incorporating user feedback from industry experts and field service engineers/SMEs.
  • โœ… Integration with existing enterprise systems and support for multiple languages?๐ŸŒ!
  • โœ… As a future scope, we plan to web scrape wikipedia pages, research papers, ebooks

The future is bright, and we can't wait to see where this journey takes us! ๐ŸŽˆ

๐Ÿง  What we learned

We learned that combining retrieval and generation techniques is like mixing peanut butter and jelly ๐Ÿฅช - they're just meant to be together! We also discovered the importance of fine-tuning the retrieval process and optimizing the model architecture, like a chef perfecting a secret recipe ๐Ÿ‘จโ€๐Ÿณ. Databricks was our trusty sous-chef ๐Ÿ‘จโ€๐Ÿณ, handling the heavy lifting of data processing and model training.

๐ŸŽ‰ Accomplishments that we're proud of

We did it! We successfully combined the retrieval and generative components into a unified RAG model that provides accurate, context-specific answers. It's like we created a mind-reading machine ๐Ÿ”ฎ that can tap into Husky's collective knowledge and deliver coherent responses. We couldn't be prouder of our brainchild! ๐Ÿ‘ถ

๐Ÿ› ๏ธ Tools Used:

  • Programming Language - SQL, PySpark, Python
  • Azure Cloud, Databricks
  • Azure Data lake Gen 2 for metastore
  • Delta Live Tables for documents storage

Built With

Share this project:

Updates