Skip to content
This repository has been archived by the owner on Jul 21, 2020. It is now read-only.

Evan, Carlie, Ryan and Audrey's Hackathon project at Horizons School of Technology

Notifications You must be signed in to change notification settings

EvanPeterson1324/horizons-hackathon-snackchat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

98 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Snackchat (Horizons Hackathon Project)

Snackchat was inspired bt Snapchat mobile application which utilizes the GoogleVision API to make distinctions between items found in photos. Much like Snapchat, our application allows users to take photos and send them to others but with a twist. Snackchat only lets the user send the photo if it recognizes the main item in the photo to be food!

Demo

alt text

Built using...

  • Google Vision API
  • React Native development with Expo
  • Prototyping/Demo with Sketch and InVision
  • MongoDB (Backend storage)
  • AWS S3 (Image cloud storage)
  • Express server: https://github.com/ryanclyde/snackchat-backend (Deployed to Heroku)
  • AJAX http requests
  • Passport authentication

Learning Points

Design Process This was the first time we've put real effort into the design process of making a 'consumer ready' product. We first made a few wireframes including every button, route, and feature we wanted the app to have. Next we made prototypes with attention to detail on the styling and colors we liked — we went with a minimal, flat UI that played with fun colors and brought about a little bit of class. This is a food app, anyway! After that, we exported the Sketch files to Invision to test out our intuitions about the user experience. Turning out styling goals into actual CSS took patience, time and nice tools like flexbox and CSS animations.

Google Vision API The API here was essential for realizing our vision of identifying what was in the pictures that our users took. What ended up being the most challenging part of it was getting it set up with the Google Vision Client, truthfully. Authentication required more than 8 keys and many tokens. G Cloud was hard to decipher. Once Google Vision was set up, we had to filter through the keywords that it gave us based on the image. If 'food' or 'beverage' was in the list, we allowed the image to be sent.

Storage None of us had ever had to store images so it was a learning curve to find out that we needed to also use AWS on top of MongoDB. MongoDB only stores JSON objects and documents up to 16mb — image files are much bigger than that. Once we got the post request of the image that the user took, we stored it in AsynchStorage then sent it out to AWS cloud storage.

Authors / Acknowledgments

  • ** Evan Peterson | Ryan Clyde | Carlie Ostrom | Audrey Setiadarma **

About

Evan, Carlie, Ryan and Audrey's Hackathon project at Horizons School of Technology

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published