Ok so I wanted to build a simple web app, that somehow would use githubarchive data. AT first I though of using the BigQuery database and it's API, however, my free quota would be over in just a day.
So, what I've done is download all 2012/2013 records from the site, and extracted them in a folder - now I have 10k+ json files than I want to use to get data for my app.
So what I want: Create a GAE (python) or Django app to visualize this data. For this, I need to convert the json files in to a database - I want to use SQLite but I'm not sure what's the best choice. I'm not sure if creating a DB is required at all, but that's what I can think of right now.
Anyone has any suggestions for me?
AppEngine is also not free beyond certain quota. I'm pretty sure that importing 10k+ Json files and then doing analytics will be well beyond the free quota.
Also, AppEngine has a very limited query language (no cross-entity queries, no joins, single inequality filter) which will prevent you doing any meaningful analytics. Go with BigQuery and swallow the cost or just do the analytics on your own machine and then export the stats to GAE for presentation.
Both GAE and BQ are a fast massively-parallel highly-available services, which you would want to use for a production applications. If you are on a budget and do not need the high-availability, go with an affordable VPS.