Pipelines for data sync of Jewish data sources to the DB of The Museum of The Jewish People
Uses the datapackage pipelines framework
This project provides pipelines that sync data from multiple external sources to the MoJP Elasticsearch DB.
Install some dependencies (the following should work on recent versions of Ubuntu / Debian)
sudo apt-get install -y python3.6 python3-pip python3.6-dev libleveldb-dev libleveldb1v5
sudo pip3 install pipenv
Install the app depepdencies
pipenv install
Activate the virtualenv
pipenv shell
Install the datapackage_pipelines_mojp package for development
pip install -e .
Get the list of available pipelines
dpp
Run a pipeline
dpp run <PIPELINE_ID>
cp .docker/docker-compose.override.yml.example.full docker-compose.override.yml
bin/docker/build_all.sh
bin/docker/start.sh
This will provide:
After every change in the code you should run bin/docker/build.sh && bin/docker/start.sh
Additional features:
docker-compose up -d kibana
docker-compose up -d adminer
Running the tests using docker
bin/docker/build_tests.sh
bin/docker/run_tests.sh
bin/docker/run_tests.sh
Make sure you have Python 3.6 in a virtualenv
bin/install.sh
cp .env.example.full .env
bin/docker/start.sh
it should work as issource .env
export DPP_DB_ENGINE=$DPP_DB_ENGINE
bin/test.sh
dpp
Clearmash is A CMS system which is used by MoJP for the MoJP own data
Clearmash exposes an API to get the data
relevant links and documentation (clearmash support site requires login)