Closed BrunoGrandePhD closed 7 years ago
brew
, apt-get
, yum
, pip
, R
et c.@sjackman Any thoughts on how to handle large files that are unfit for version control? For instance, a few weeks ago, I had a 120-MB file that I couldn't commit to GitHub (limit is max. 100 MB). Git LFS and git-annex would be logical solutions to this problem, but I've also read (sorry, no links offhand) that they pose some problems. Notably, Git LFS disables forking on a repository, and cloning a repository is now more complex for new users who aren't familiar with Git LFS.
Git LFS disables forking on a repository
Ugh. That's terrible. I didn't know that about LFS. My only suggestion is to archive the data online somewhere appropriate, and to include the URL and SHA-256 in the build script.
Description
Do you like the idea of reproducible science, but don't know where to start? Have you been told that you should use Git, but never got around to learning it? Join this workshop to learn about various techniques (e.g. R projects, Git and R Markdown) that will help your science be more reproducible. We will start a simple data analysis project from scratch and build it up while following best practices for reproducibility.
Time and Place
Where: Simon Fraser University, Burnaby Campus, Library Research Commons When:
Monday, August 15th, 10:30-12:30 am/ Delayed until fall⟶ Note: This is a two-hour workshop.
Registration
REGISTER HERE
Call for Tips/Links
They say the best way to learn something is to teach it. I'll be doing my research, but I still welcome any feedback I can get on the topic. Feel free to comment below with tips or links on best practices for reproducible data analyses. Thanks in advance!