preprocessing-sfa is an Enduro preprocessing workflow for SFA SIPs.
The preprocessing workers need to share the filesystem with Enduro's a3m or Archivematica workers. They must be connected to the same Temporal server and related to each other with the namespace, task queue and workflow name.
The required configuration for the preprocessing worker:
debug = false
verbosity = 0
sharedPath = "/home/enduro/preprocessing"
[temporal]
address = "temporal.enduro-sdps:7233"
namespace = "default"
taskQueue = "preprocessing"
workflowName = "preprocessing"
[worker]
maxConcurrentSessions = 1
Optional BagIt bag configuration:
[bagit]
checksumAlgorithm = "md5"
The preprocessing section for Enduro's configuration:
[preprocessing]
enabled = true
extract = false
sharedPath = "/home/enduro/preprocessing"
[preprocessing.temporal]
namespace = "default"
taskQueue = "preprocessing"
workflowName = "preprocessing"
This project uses Tilt to set up a local environment building the Docker images in a Kubernetes cluster. It has been tested with k3d, Minikube and Kind.
A local Kubernetes cluster:
It can run with other solutions like Microk8s or Docker for Desktop/Mac and even against remote clusters, check Tilt's Choosing a Local Dev Cluster and Install documentation for more information to install these requirements.
Additionally, follow the Manage Docker as a non-root user post-install guide
so that you don’t have to run Tilt with sudo
. Note that managing Docker as a
non-root user is different from running the docker daemon as a non-root user
(rootless).
While we run the services inside a Kubernetes cluster we recomend installing Go and other tools locally to ease the development process.
Start a local Kubernetes cluster with a local registry. For example, with k3d:
k3d cluster create preprocessing --registry-create sdps-registry
Or using an existing registry:
k3d cluster create preprocessing --registry-use sdps-registry
Make sure kubectl is available and configured to use that cluster:
kubectl config view
Clone this repository and move into its folder if you have not done that previously:
git clone git@github.com:artefactual-sdps/preprocessing-sfa.git
cd preprocessing-sfa
Bring up the environment:
tilt up
While the Docker images are built/downloaded and the Kubernetes resources are
created, hit space
to open the Tilt UI in your browser. Check the Tilt UI
documentation to learn more about it.
Tilt, by default, will watch for file changes in the project folder and it will
sync those changes, rebuild the Docker images and recreate the resources when
necessary. However, we have disabled auto-load within the Tiltfile to reduce
the use of hardware resources. There are refresh buttons on each resource in the
Tilt UI that allow triggering manual updates and re-executing jobs and local
resources. You can also set the trigger_mode
env string to TRIGGER_MODE_AUTO
within your local .tilt.env
file to override this change and enable auto mode.
Run ctrl-c
on the terminal where tilt up
is running and stop the cluster
with:
k3d cluster stop preprocessing
To start the environment again:
k3d cluster start preprocessing
tilt up
Check the Tilt UI helpers below to just flush the existing data.
To remove the resources created by Tilt in the cluster, execute:
tilt down
Note that it will take some time to delete the persistent volumes when you
run tilt down
and flushing the existing data does not delete the cluster.
To delete the volumes immediately, you can delete the cluster.
Deleting the cluster will remove all the resources immediatly, deleting cluster container from the host. With k3d, run:
k3d cluster delete preprocessing
A few configuration options can be changed by having a .tilt.env
file
located in the root of the project. Example:
TRIGGER_MODE_AUTO=true
Enables live updates on code changes for the preprocessing worker.
In the Tilt UI header there is a cloud icon/button that can trigger the preprocessing workflow. Click the caret to set the path to a file/directory in the host, then click the cloud icon to trigger the workflow.
Also in the Tilt UI header, click the trash button to flush the existing data. This will recreate the MySQL databases and restart the required resources.
The Makefile provides developer utility scripts via command line make
tasks.
Running make
with no arguments (or make help
) prints the help message.
Dependencies are downloaded automatically.
The debug mode produces more output, including the commands executed. E.g.:
$ make env DBG_MAKEFILE=1
Makefile:10: ***** starting Makefile for goal(s) "env"
Makefile:11: ***** Fri 10 Nov 2023 11:16:16 AM CET
go env
GO111MODULE=''
GOARCH='amd64'
...