These Python scripts can be used to collect book reviews and metadata from Goodreads.
We were motivated to develop this Goodreads Scraper because the Goodreads API is difficult to work with and does not provide access to the full text of reviews. The Goodreads Scraper instead uses the web scraping libraries Beautiful Soup and Selenium to collect data.
We used this Goodreads Scraper to collect data for our article, "The Goodreads ‘Classics’: A Computational Study of Readers, Amazon, and Crowdsourced Amateur Criticism." To allow others to reproduce (approximately) the data we used in the essay, we include a file with 144 Goodreads book IDs for the 144 classics that we analyzed (goodreads_classics.txt
). You can use these IDs to collect corresponding reviews and metadata with the Goodreads Scraper as described below.
Note: Updates to the Goodreads website may break this code. We don't guarantee that the scraper will continue to work in the future, but feel free to post an issue if you run into a problem.
Goodreads recently updated their book pages with a new layout. We've heard that our scraper is still functioning, but we haven't fully tested or updated the scraper to account for these changes.
To run these scripts, you will need Python 3.
You will also need the following Python libraries:
You can install these Python libraries by running pip install -r requirements.txt
Finally, you will need a web browser — either Chrome or Firefox. We have found that the Goodreads Scraper tends to function better with Firefox.
We recommend running these Python scripts from the command line, as the usage instructions below describe. However, we have also created a Jupyter notebook tutorial that demonstrates how to use the Goodreads Scraper scripts. Please note that these scripts may not work consistently from a Jupyter notebook environment and that the tutorial is mostly intended for demonstration purposes.
You can use the Python script get_books.py
to collect metadata about books on Goodreads, such as the total number of Goodreads reviews and ratings, average Goodreads rating, and most common Goodreads "shelves" for each book.
This script takes as input a list of book IDs, stored in a plain text file with one book ID per line. Book IDs are unique to Goodreads and can be found at the end of a book's URL. For example, the book ID for Little Women (https://www.goodreads.com/book/show/1934.Little_Women) is 1934.Little_Women
.
This script outputs a JSON file for each book with the following information:
This script also outputs an aggregated JSON file with information about all the books that have been scraped. To output an aggregated CSV file in addition to a JSON file, use the flag --format CSV
.
python get_books.py --book_ids_path your_file_path --output_directory_path your_directory_path --format JSON (default) or CSV
python get_books.py --book_ids_path most_popular_classics.txt --output_directory_path goodreads_project/classic_book_metadata --format CSV
You can use the Python script get_reviews.py
to collect reviews and review metadata about books on Goodreads, including the text of the review, star rating, username of the reviewer, number of likes, and categories or "shelves" that the user has tagged for the book.
This script takes as input a list of book IDs, stored in a plain text file with one book ID per line. Book IDs are unique to Goodreads and can be found at the end of a book's URL. For example, the book ID for Little Women (https://www.goodreads.com/book/show/1934.Little_Women) is 1934.Little_Women
.
This script outputs a JSON file for each book with the following information:
This script also outputs an aggregated JSON file with information about all the reviews for all the books that have been scraped. To output an aggregated CSV file in addition to a JSON file, use the flag --format CSV
.
Goodreads only allows the first 10 pages of reviews to be shown for each book. There are 30 reviews per page, so you should expect a maximum of 300 reviews per book. By default, the reviews are sorted by their popularity. They can also be sorted chronologically to show either the newest or oldest reviews.
We also select a filter to only show English language reviews.
python get_reviews.py --book_ids_path your_file_path --output_directory_path your_directory_path --browser your_browser_name --sort_order your_sort_order --rating_filter your_rating_filter --format JSON (default) or CSV
sort_order
can be set to default
,newest
or oldest
.
rating_filter
can be omitted or set to any number in the range 1-5.
browser
can be set to chrome
or firefox
.
format
can be set to JSON
(default) or CSV
.
python get_reviews.py --book_ids_path most_popular_classics.txt --output_directory_path goodreads_project/classic_book_reviews --sort_order default --rating_filter 5 --browser chrome
You can run the provided test script to check that everything is working correctly.
./test_scripts.sh
This will create a directory called test-output
in which you'll find the scraped books and reviews.
This code is written by Maria Antoniak and Melanie Walsh. The code is licensed under a GNU General Public License v3.0.
If you use this scraper, we'd love to hear about your project and how you use the code.
If you use this scraper as part of an academic publication, you can credit us by citing the following paper.
Walsh, Melanie, and Maria Antoniak. "The Goodreads ‘Classics’: A Computational Study of Readers, Amazon, and Crowdsourced Amateur Criticism." Journal of Cultural Analytics 4 (2021): 243-287.
We used a function written by Omar Einea, licensed under GPL v3.0, for the Goodreads review sorting.