There are times when we need to scrape some data from a website to use in our analyses. In a perfect world, the provider of the data would provide a csv or json download but let’s face it… we do not live in a perfect world and often data that is posted to the internet is done by people who really do not care about subseqent use by others (else they would have made it easy to use rather than just showing it).
This is a moderatly interesting proposal. The main things that I’ve been dealing with are:
Getting crap out of WordPress. My old site is being hosted by an outside provider and I thought I’d pick up a new static site using GitLab since they allow private repositories without paying.
OK, so there is a bit of a circular firing squad going on in some of my R installs with ggplot2. Apparently, you can get various CRAN/Github versions out of sync and a whole host of different. Here is how it started: