This is a very hastily done brainstorm. As part of the Revitalizing Monographs workshops various premises and thought experiments were raised that suggested that part of what needs to be done to "revitalize" monography is to point the broader community to tools, processes, and methods that already exist. We don't mean to imply that things are "solved" in any of these areas. We do mean to imply that there are numerous, often very good, options out there for getting the core data in a Monograph into others hands/machines, and that some of these options adequately address requests being made of monographers. There are also emerging technologies that seek to serve Monographers explicitly that people may not be aware of.
This is an opinionated list. It's trying to make a particular point (and raise others) rather than be comprehensive/correct etc.
The content here is gathered in such a format that it could, maybe, be summarized in a 5-6 minute lightning talk.
You can contribute to this by either opening an issue or making a pull request. All errors, misinterpretations, etc. are Matt Yoder's fault, not the collaborating contributors.
CC 0.
Each "My X" section/slide represents a type of data in a monograph and has up to three categories: 1) (Do) Now - You can do this now! 2) (Do) Soon - Sensu the Workshop's "I can imagine this technology existing already", not "feature available tomorrow" 3) (To) do - Quick ideas/opinions/comments about missing bits.
TODO? - At the end of this section a series of hyperlinks reference the issues raised at the workshop as recorded on the "stickies" there.
Get an ORCID id, get your collaborators ORCIDs. Add ORCIDs to your data (e.g. in recordedByID
or identifiedByID
). Get a Wikidata Q identifier for your people, for dead people who collected the specimens you looked at. Spam these IDs everywhere you archive other data (e.g. Figshare, Zenodo, Publishers, all take ORCIDs now). Publish your data with ORCIDs to GBIF, watch people find your data by looking you up at GBIF. You're a primary curator/producer, so Bionomia has less impact for you during production of the monograph, i.e. by definition you're re-examining specimens and their data, but when your data get to GBIF, then your data will be visible in platforms like Bionomia, i.e. more accessible. Use tools like TaxonWorks to reference People as things, not rows of data while you compile your monograph.
Publish your People to Wikidata via tools like TaxonWorks.
Ensure your physical specimens have catalog numbers. Share your specimens with GBIF via DwC format. Pensoft will help do this. There are many other routes including publishing your dataset on Zenodo, GitHub, IPTs, etc. and pointing GBIF there. Numerous platforms exemplify tools that facilitate the pre-aggregation and curation of these data (e.g. Specify, TaxonWorks, Symbiota, Open Refine, Arctos). All of these data form the core of the "Digital Specimen". You, as a monographer don't have to worry about that downstream job, just remember: 1) Catalog numbers on specimens; 2) Recommended bonus: UUIds for the occurrence going to GBIF as well.
Better rendering of specimens in human readable format.
Relatively speaking this is very well worked out.
Add your data to Zoobank, Pensoft will do it all for you, you'll get LSIDs for names by doing this. You can do it yourself if you want. Or don't. Share your nomenclature to GBIF via a DwC Checklist. Share your nomenclature to the Catalog of Life via basically any format you want (assuming its a monograph with global coverage). Share your data to the World via the CoLDP format (TaxonWorks, WORMS, many others). Share your ridiculously semantic NOMEN referencing nomenclature via a public TaxonWorks API.
Load your tree, pre-publication even, to Open Tree.
Just put some sort of table on a Github pages (or Google Drive, or anywhere that table is online) and whisper in the ear of Globi, you're good.
The text description. If it's published in Pensoft it's been processed by Plazi and isolatable via Pensoft and Plazis tools. You can also publish NeXML, Nexus, TNT, CSV versions of your data your companion GitHub repository. You can share your Observations (matrix, phylogenetic or descriptive) via a TaxonWorks JSON serving endpoint. You can do crazy things like make Manchester syntax statements with rediculous OWL references (e.g. go to the lengths to produce RDF triples). You can use old tools like LUCID and others that will spit out SDD format. You could co-opt Mesquite or other matrix building software to manage your states in a matrix. You can use a simple script to compose your telegraphic sections (tools like 'mx', vSysLab, etc. have been doing this for over a decade, tools like IntKey have been doing this for decades). Create RC5/21 assertions between your taxa, share the corresponding CSV on your companion Git repository.
More flash, not much more substance.
MorphDBase2 levels of instance based assertions.
Scientists have been sending data to Genbank for a long time. Most importantly, you're a monographer and know that unique-identifier based vouchering, or the use of UUIDs when vouchering is not possible, is the foundation of your monograph, so you include references to those in your uploads whenever possible. You can also send your data to BOLD.
OK, this is awful. But wait- Zenodo, Figshare, MorphoSource, Morphobank! Upload your plates. Get DOIs. Share all your images, and their metadata via a TaxonWorks API.
Cleverly point Internet Archive at your images.
Use IPFS peer-to-peer to share your images bit-torrent-esqe style.
Use a Juptyer notebook if crunching numbers. Add that to your monographs companion git repo. Publish your generalized notes as a Git repo. Example. You can also create your own citable protocols at Protocols.io.
Did you know that adding a repository to Github is perhaps the single best ultra long term archive of your data at the moment? Zip everything, compute a SHA for the zip, reference the SHA in your publication. Put the Zip and the SHA "out there". Also, make crude webpages using free Github pages, point the Internet Archive at them.
Tools like TaxonWorks could package your data from all the steps, sans images, and shove them on a Git repo or Publisher for you.
Git + images (or large genomic datasets) doesn't work so well.
Crossreference an OBO foundry ontology. Contribute your data to a growing number of anatomical ontologies. Get your vertebrate data in Phenoscape's hands. Create a "URI table" where possible. Use ImageJ and other tools to auto-measure your anatomy and provide tables of data (e.g. Ants, Odontates, Fish) to your companion Git repo. Data matrices can be archived as NEXUS files into Dryad.
Moaaar anatomy ontologies. Semantic labelling of images.
Does your monograph contain stuff like this? Then it can be shared in places like this.