Skip to content

mfriebel/tweet_bot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Tweets Collection and Analysis Pipeline

This project implements a data pipeline using Docker. Tweets are streamed about sustainability are streamed via Tweepy Listener and stored in a MongoDB database. The ETL job performs live sentiment analysis (using VADER) on the stored tweets and loads them with the according score into a PostgreSQL database. In the end tweets with most positive sentiment are posted on Slack using a Webhook.

Pipeline

Docker Compose necessities

Setting up local environmental variables

Changing streaming filter

The file get_tweets.py contains the Tweepy Tweets Listener. The topic filter is found at the end of the file: stream.filter(track=['sustainable'], languages=['en'])

About

This project implements a data pipeline using Docker to extract, transform (sentiment analysis) and load twitter posts and finally post them on Slack.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors