Fabio Rosado

My time is divided between flying and coding

Run script to run all the scrapy spiders and check how many items we can get (around 600)

Change export encoding to utf-8 for right encoding in json when scrapper is run

Check each url and split url from mail rss scrapper to get the right categories and source

Use nltk.SentimentIntensityAnalysis to add another layer to the classify method

Update the classifier code to use old twitter one since the new one is way too slow!

Create helper function to check if date from article is from todays date and return bool

Test new classifier with json file and check how quick it can classify compared with actual one

Shipping live on Shipstreams! https://shipstreams.com/FabioRosado

Replace classifier with old twitter classifier - got good results and it’s blazing fast!

Improve classification speed by moving loading of the vocabulary and classifier into the classifier class

Create script to run all the scrapers at once and get the data into the same file