* [MongoDB](https://www.mongodb.com/) and [PyMongo](https://api.mongodb.com/python/current/) for data persistence
* [Beautiful Soup](https://www.crummy.com/software/BeautifulSoup/) for parsing HTML
-(You may want to install [Anaconda] to get most of these in one bundle.)
+(You may want to install [Anaconda](https://anaconda.org/) to get most of these in one bundle.)
## Data sources
Note that the NRC sentiment lexicon isn't available for download, but you'll have to email the creator.
-You can obtain a cache of what I gathered in the `data` directory of this repo. Use `mongorestore` to recover the data.
+You can obtain a cache of what I gathered in the `dump` directory of this repo. Use `mongorestore` to recover the data.
## Use
The simplest and best-documented notebooks are the [Beatles vs Stones: gather data](beatles-vs-stones-gather-data.ipynb) and [Beatles vs Stones: analysis](beatles-vs-stones-analysis.ipynb) notebooks. Once you've seen those, take a look at [generic data gathering](multi-artist-gather-data.ipynb) and [generic data analysis](multi-artist-analysis.ipynb) to do the wider analysis. In particular, the [generic data analysis](multi-artist-analysis.ipynb) notebook does the convex hull calculation that generated the figures in the presentation.
-You'll also need to run through the [Tag lyrics with emotions](/tag-lyrics-with-emotions.ipynb) notebook if you're using the NRC sentiment corpus.
\ No newline at end of file
+You'll also need to run through the [Tag lyrics with emotions](/tag-lyrics-with-emotions.ipynb) notebook if you're using the NRC sentiment corpus.