Kuwala is the data workspace for BI analysts and engineers enabling you to build powerful analytics workflows together. We are set out to bring state-of-the-art data engineering tools you love, such as Airbyte, dbt and Prefect together in one intuitive interface built with React Flow.
Do you want to discuss your first contribution, want to learn more in general, or discuss your specific use-case for Kuwala? Just book a digital coffee session with the core team here.
Kuwala stands for extendability, reproducibility, and enablement. Small data teams build data products fastly and collaboratively. Analysts and engineers stay with their strengths. Kuwala is the tool that makes it possible to keep a data project within scope while having fun again.
Currently we support the following databases and data warehouses
[x] Snowflake
For connecting and loading all your tooling data into a data warehouse, we are integrating with Airbyte connectors. For everything related to third-party data, such as POI and demographics data, we are building separate data pipelines.
To apply transformations on your data, we are integrating dbt which is running on top of your data warehouses. Engineers can easily create dbt models and make them reusable for the frontend. We have already a catalog of several transformations that you can use on the canvas. The complete documentation can be found here: https://docs.kuwala.io/
We are going to include open-source data science and AI models as blocks (e.g., Meta's Robyn Marketing Mix Modeling).
You can easily connect your preferred visualization tool and connect it to a saved table on the canvas in the future. We will make the results exportable to Google Sheets and also available in a Medium-style markdown editor.
With the canvas you can connect to your data warehouse and start building data pipelines. To start the canvas, simply run the following command from inside the root directory:
docker-compose --profile kuwala up
Now open http://localhost:3000 in your browser, and you are good to go. 🚀
We currently have five pipelines for different third-party data sources which can easily be imported into a Postgres database. The following pipelines are integrated:
To use Kuwala's components, such as the data pipelines or the Jupyter environment, individually, please refer to the
instructions under /kuwala
.
Every new issue, question, or comment is a contribution and very welcome! This project lives from your feedback and involvement!
The best first step to get involved is to join the Kuwala Community on Slack. There we discuss everything related to our roadmap, development, and support.
Please refer to our contribution guidelines for further information on how to get involved.
Link | Description |
---|---|
Blog | Read all our blog articles related to the stuff we are doing here. |
Join Slack | Our Slack channel with over 250 data engineers and many discussions. |
Jupyter notebook - Popularity correlation | Open a Jupyter notebook on Binder and merge external popularity data with Uber traversals by making use of convenient dbt functions. |
Podcast | Listen to our community podcast and maybe join us on the next show. |
Digital coffee break | Are you looking for new inspiring tech talks? Book a digital coffee chit-chat with one member of the core team. |
Our roadmap | See our upcoming milestones and sprint planing. |
Contribution guidelines | Further information on how to get involved. |