We really like food, but wouldn't want eating it coming at a risk. Being able to understand the food you're about to eat and whether or not you're allergic to it with just a simple picture is simple and fun.

What it does

The user enters their allergies into the app, which can be used to take pictures of food and pull out all the ingredients. It will be able to warn users if ingredients they are allergic to is in the food, while also telling the user information about the food (vegan, vegetarian, etc.)

How we built it

We used react native through to build the app, along with flask. Detected food using google cloud vision API, and used python with the Edamam API to scrape all ingredients and info about the food.

Challenges we ran into

We ran into many challenges, such as sticking everything together and debugging the many errors we encountered working with react native.

Accomplishments that we're proud of

We managed to integrate different APIs together to create something unique and cool.

What we learned

Got better at react native and development of mobile apps.

What's next for AllerVision

Have a nicer GUI, and improve the analytics it provides on food.

Share this project: