Open arthursunbao opened 6 years ago
If you have error during make with this error: "pandoc version 1.12.3 or higher is required and was not found (R shiny)"
Please try to use export RSTUDIO_PANDOC=/Applications/RStudio.app/Contents/MacOS/pandoc for macOS and export RSTUDIO_PANDOC=/usr/lib/rstudio/bin/pandoc for Ubuntu in terminal (https://stackoverflow.com/questions/28432607/pandoc-version-1-12-3-or-higher-is-required-and-was-not-found-r-shiny/29710643#29710643)
Your main README looks great, but there's no link to homework 8, 9, or 10 there. You may want to add those links to your list!
Your readme for HW10 is very thorough! Tons of content in your progress report!
It's great that you decided to do both "Make API queries by hand" AND "Scrape the web" assignments.
There's an inner README for this subsection, which is a nice orientation to the work you have done here.
I would suggest using a more descriptive title for the omdb_dataset_download_show.md
file. The filename and title don't have to be the same!
It would be best if you masked the output of all of your library
function calls by updating R and/or using the function suppressPackageStartupMessages()
. The text warnings are not important for your data analysis but take up a lot of space.
It's great that you incorporated a makefile
into your data analysis pipeline here!
The code in your R script is easy to follow. You demonstrate a strong understanding of httr
usage for downloading data from API's.
You provide plenty of explanatory text and comments, making it very easy to follow along with your code.
Your plots and tables demonstrate that you have succeeded in the task you set out to accomplish here.
To keep this subdirectory less cluttered, I always find it's best practice to put all images and data into an inner subdirectory.
Your plot.R
script here contains two lines that will make tables using knitr::kable
. Those lines should be moved to your .rmd
file as R scripts are usually not executed interactively within Rstudio.
You successfully saved a couple of tidy data frames that are ready for some downstream data analysis. You even automated the pipeline for doing this with makefile
so that you can easily go back and easily get different data later.
Once again you provide an inner README file in this subdirectory which helps to orient the reader.
Again, you should use a better title for scrape_download_show.md
and hide the outputs of the library
functions since they are not really important.
You were successfully able to extract the links from a list of links at http://www.bestplaces.net/find/county.aspx?counties=mo
and save that to a CSV file.
You were able to give an example of extracting crime info for a particular county using one of your links.
It would have been better if you were able to construct a dataframe of some data using your list of links, and perhaps generate some interesting plots. Then save the data frame containing actual data instead of links.
Overall: You did very well. Clearly a lot of work put into this assignment!
Cheers,
Hayden
Method Chosen: httr & scrape Gets data from web: Yes Clean and tidy data frame: Yes Basic Data Exploration: Yes Bonus: modify url (request different titles) Report/Reflection: Yes
Comments:
Your mark will be distributed later. If you would like more feedback, please feel free to message me on slack.
https://github.com/arthursunbao/STAT545-Homework/tree/master/hw10