Inspiration

Hallucinations are important to identify. inspect-LLM does that in an explainable and quantifiable way. Even in normal ML evaluation, people only care about the high level metrics, and don't really understand where model A and model B differ, even if their overall performance is the same number.

What it does

TruthfulQA -> Cohere Chat -> unbiased 3rd party LLMs for hallucination detection -> graph to demonstrate failures (subset + count, coloured by performance metric).

How I built it

APIs + mathplotlib

Challenges I ran into

Volleyball tournament in the afternoon.

Accomplishments that I'm proud of

New insights

What I learned

Cohere chat needs to work on prompt adherence

What's next for inspect-LLM

Nothing - just a learning exercise. Those with resources might consider this as an internal tool.

Built With

Share this project:

Updates