Open annakrystalli opened 6 years ago
This sounds great! Might also have tie-ins to #41 ?
To throw a bit more into the mix, I think Google's Dataset description would be worth some exploring too, both because 'hey, if Google supports it, that could definitely help with discovery' and also because I think it's a pretty light-weight / user-friendly approach (says my sample size of n=1
).
@sckott and I have tossed around the idea of using this to get a better picture of exactly what datasets rOpenSci packages can access, e.g. movebank example, GBIF.
I think it would be nice to think of a prototype package that could help generate the basic bits (i.e. bibliographic info, spatial & temporal coverage of the data, data type), and also help you quickly and easily do something cool with that metadata (i.e. search interface, indicate the region on a map, period in time, etc). I think getting immediate value to the researcher back out is an important missing piece of metadata descriptions; adding metadata often feels both way too painful and too altruistic.
A nice feature of Google's Dataset description is that it's JSON, so basically equivalent to an R list object already, but since it's all in the https://schema.org context, we get links to ontologies "for free" (such as the W3C's DCAT) because these maps are already built in OWL in the schema.org definitions. Other cases one would have to map more 'manually' but that might be easier anyway! We have some very sketchy outlines for working with Schema.org in https://github.com/ropenscilabs/datasauce
Ditto--This sounds wonderful! Cheers @annakrystalli for trying to integrate metadata standards into budding ecologists' workflows -- & may there be delicious datasauce soon!
I think it'd be great to integrate #52 here -- Should I [try to] merge it, or is that jumping any guns?
And I agree with @cboetig on relevance to #41. I'm tempted to think of cleaning/standardizing/analysis/publication as distinct/distinguishable stages in data-handling processes, but I don't want to oversimplify if folks have other thoughts on life-cycles/relationships in/around their datasets? (Including 'fitness for [which] use/s'...) Building around the research compendium/analysis best practices mentioned here + in #64 sounds like it would help wrap heads around the issues.
@aurielfournier & @annakrystalli - "Curating Research Data" includes an ecology-related case study, along with others from a variety of fields if those might be handy or already familiar to some other folks here.
JSON-LD sounds pretty sweet--Anyone happen to know how long those have been in the works? (or how long I've been under this rock?!)
👍 on this and 👍 also on "doing something cool"! I also think that's where people fall off the metadata wagon.
Good example dataset to document https://knb.ecoinformatics.org/#view/urn:uuid:0255141e-b0b0-4dbe-8586-8838f2ac9048
link to piratenpad: https://piratenpad.de/p/dataspice
Link to project repo: https://github.com/ropenscilabs/dataspice
As a results of teaching Research Data Management to first year primarily biology and ecology PhD students using R for the passed 4 yrs, I’ve been thinking a lot about metadata. In particular what the most basic level we teach students should be and what sort of files and formats to use.
I tried to teach them to create Ecological Metadata Language objects (EML) using pkg
eml2
last year but I wouldn’t consider my attempt successful. It seemed too much for them at this point in their studies and not critical enough to spend more of the limited time we had together on it.But what I’ve tried to do since and feel I wanted them to ultimately get out of the session was to learn to prepare metadata files that are:
csvy
,eml2
etc) that they can get to grips with further down the line, as and when it’s required.So I’ve been thinking about R functionality that could help with both the task of creating said basic files and establishing best metadata management practice. Some of the features I’ve envisaged would help guide or semi-/automate some of the requirements identified in rOpenSci Analysis Best Practice Guidelines as still being human powered, eg.
I made a start putting some useful functionality together in nascent package
metadatar
(see draft README for quick demo of some of the functions) that I used successfully for teaching. The main function creates a metadata table for a given dataset, with a row for each column of the original dataset. The basic metadata fields are based on the EML attribute table and the function tries to automatically populate some of the fields but the rest need to be entered manually still. However, it would be great if this stage could be integrated with the excellent suggestions in issue #52 and perhaps through #68? (I need to get my head round CEDAR a bit more to understand how this might fit in).To show students how keeping good metadata about variables can be useful in analyses, there’s also a function to extract more informative variable descriptions and units for plotting and presenting results in tables from the metadata table.
Would love to hear other people’s thoughts on this and their own current and ideal approaches to these issues and perhaps even work on developing some more tools for this during the unconf. It feels like a topic that could perhaps be integrated with issue #52?
Summary of proposal