Please describe the problem you'd like to be solved.
The Archivematica sample data repo has grown organically over the course of the project, added to and updated as needed. The repo currently has five top-level directories:
OPF format-corpus @ 5ddb1a2 (An openly-licensed corpus of small example files, covering a wide range of formats and creation tools maintained by the OPF)
SampleTransfers - in-house sampledata created to test features and functionality
TestTransfers - in-house sampledata created to test failure/error conditions
createtransfers - a tool for creating large/weird test corpuses
training-resources - in-house sampledata used for training sessions by Artefactual staff in the past
There is a lot of sample data here, but it's not well documented (in the sense that there is no record of the conditions under which a transfer should be used and what results the user should expect). The transfer names themselves do not always clearly indicate what the transfer is meant to do/show. The distinction between SampleTransfers and TestTransfers is not clear, especially when the transfer name is not descriptive. training-resources was a nice idea but is maybe not the most easily-maintained thing. It's pretty easy for a user to pick up a transfer that is meant to fail without realizing it.
As a more general note, many of the sample transfers could be more robust and/or more interesting to look at - we use these transfers frequently, sometimes it gets boring seeing the same thing all the time!
Describe the solution you'd like to see implemented.
I think addressing a couple of things would really help clarify how the transfers should be used:
Reorganize the sample data repository into clear functional categories (i.e. FeatureTransfers for testing specific features, ErrorTransfers for transfers that should error out in some way, etc.)
Make sure transfer names are descriptive
Create a list of each transfer, its processing requirements, and the expected outcomes
Move the automated-tests folder up to the top level and ensure that transfers used for AMAUAT tests are located there (this gives them a bit more visibility and also makes it less likely that someone will accidentally change the transfer)
Describe alternatives you've considered.
I guess we could leave it as is?
Please describe the problem you'd like to be solved. The Archivematica sample data repo has grown organically over the course of the project, added to and updated as needed. The repo currently has five top-level directories:
There is a lot of sample data here, but it's not well documented (in the sense that there is no record of the conditions under which a transfer should be used and what results the user should expect). The transfer names themselves do not always clearly indicate what the transfer is meant to do/show. The distinction between
SampleTransfers
andTestTransfers
is not clear, especially when the transfer name is not descriptive.training-resources
was a nice idea but is maybe not the most easily-maintained thing. It's pretty easy for a user to pick up a transfer that is meant to fail without realizing it.As a more general note, many of the sample transfers could be more robust and/or more interesting to look at - we use these transfers frequently, sometimes it gets boring seeing the same thing all the time!
Describe the solution you'd like to see implemented. I think addressing a couple of things would really help clarify how the transfers should be used:
FeatureTransfers
for testing specific features,ErrorTransfers
for transfers that should error out in some way, etc.)automated-tests
folder up to the top level and ensure that transfers used for AMAUAT tests are located there (this gives them a bit more visibility and also makes it less likely that someone will accidentally change the transfer)Describe alternatives you've considered. I guess we could leave it as is?