As part of a research project, I have a script that consumes tweets from twitter into a locally hosted mongodb database:
import json
import pymongo
import tweepy
consumer_key = ""
consumer_secret = ""
access_key = ""
access_secret = ""
auth = tweepy.OAuthHandler(consumer_key, consumer_secret)
auth.set_access_token(access_key, access_secret)
api = tweepy.API(auth)
class CustomStreamListener(tweepy.StreamListener):
def __init__(self, api):
self.api = api
super(tweepy.StreamListener, self).__init__()
self.db = pymongo.MongoClient().test
def on_data(self, tweet):
self.db.tweets.insert(json.loads(tweet))
def on_error(self, status_code):
return True # Don't kill the stream
def on_timeout(self):
return True # Don't kill the stream
sapi = tweepy.streaming.Stream(auth, CustomStreamListener(api))
sapi.filter(track=['snowden'])
To improve uptime, I would like to do two things: i) run this script remotely, and ii) store the consumed tweets in the cloud. However, being completely new to all things programming, I am lost as to what I should do to achieve my goals. What are my next steps? What is the "path of least resistance" to uptime?
Heroku is a cloud platform supports Python and MongoDB which I'd recommend you use. This link provides a working reference on how to do this.
Here are another couple of links to help you out:
1) Python database WITHOUT using Django (for Heroku)
2) How can I use the mongolab add-on to Heroku from python?
Hope this helps!