bcodell / dbt-activity-schema

A dbt package with a POC implementation of an interface to query activity streams that adhere to the Activity Schema 2.0 spec.
Apache License 2.0
12 stars 2 forks source link

dbt-activity-schema

A dbt package to build and maintain Activity Schema data pipelines and build datasets from Activity Streams. Assumes some working knowledge of the Activity Schema Spec.

Sponsors

Big thanks to LocalStack for providing PRO licenses to the development team! Their support enables compatibility testing for Snowflake in this project's CI suite.

Motivation

After contributing to multiple dbt projects that support Activity Schema modeling in dbt, two observations became apparent:

  1. dbt is a reasonable environment for building and maintaining data pipelines that adhere to the Activity Schema spec.
  2. A macro-based UX for generating ad hoc datasets from Activity Stream tables for analysis development left a lot to be desired.

This project exists for two reasons - to make it easier to operate an Activity Schema, and to offer a preliminary proposal for a standard method for deriving datasets from stream-like tables.

The benefits for operating an Activity Schema are obvious - the data development workflow for this methodology is highly repetitive, and dbt's macro interface is good enough to abstract away most of that redundancy without too much pain.

However, when macros require dozens of input arguments, they become painful to use, and this case applies particularly to building datasets from Activity Streams. For one, developers need to memorize the input arguments for the macro and their corresponding formats. Is the input a string? A dbt ref? Another macro? Rendered or unrendered? This challenge compounds on the existing need for developers to learn the semantics of Activity Schema modeling even before leveraging a tool to build it.

Second, all versions of a macro to produce a dataset from an Activity Stream are particularly verbose - declaring a stream, a primary activity, one or more joined activities, their relationships to the primary activity, columns to include, aliases and aggregations to apply to each column - in addition to the tediousness of defining all of these attributes in a macro, the code gets less and less readable as the dataset is expanded (i.e. more columns are added).

But ultimately, there's an emotional component as well. Data professionals enjoy writing SQL. They're familiar with it. They're good at it. It's easy to achieve a flow state while writing it. But the UX of a dbt macro is decidedly un-SQL-like. Developers embrace macros when it allows them to make their SQL more DRY - but to provide a solution that entirely strips away any SQL writing from their work, particularly given the aforementioned challenges, feels like one that is missing the mark.

Given all of these shortcomings, It wouldn't be surprising if developers could write bespoke SQL against an Activity Stream table faster than they can populate a complex and unfamiliar dbt macro. However, writing that SQL from scratch is cumbersome - lots of redundant self-joins and filters - and cumbersome enough to give developers pause from adopting this modeling paradigm.

This project offers a middle ground. With aql, developers can derive their datasets from Activity Streams in a way that is less tedious than writing traditional SQL, easier to learn and read than a convoluted macro, and above all, it feels like writing SQL (with a few new keywords).

That said, proposing (yet) a(nother) new DSL is treacherous territory; there's plenty of literature in staunch opposition of anything other than SQL (see one oldie-but-goodie example here). And that's fair - SQL is well-known and malleable. But as we've highlighted, writing SQL against an activity stream in this modeling paradigm is tedious, redundant, and parameterizable. So this package offers a macro-based interface for generating queries in line with the Activity Schema spec. And for the purists reading this, writing bespoke SQL is also option! That's what dbt was made for anyway 😅

Inspiration

Numerous projects influenced the design of this one. Such projects include:

Implementation Design Decisions

At its core, this project is effectively a proposal for for an alternative data modeling standard with tooling tailored to the methodology. In order to obtain meaningful feedback and iterate on such a proposal, it needs to meet its prospective users where they already live. In the analytics world, that place is dbt.

For the scaffolding to support activity definitions, building streams, incremental updates, identity resolution, and all of the other needs of building an Activity Schema data pipeline, a dbt package is the logical choice. Developers mostly write sql, and call the corresponding macros to make their code DRYer (more DRY? 🤷‍♂️). What's less certain is the implementation strategy for the experimental aql.

On that subject, there were a couple paths forward for building this project. One is to build a standalone package, with integrations for use in other dialects, languages, and frameworks. Prql is a great example of a similar project. But while admirable implementations exist to integrate such projects with dbt, a standard integration pattern has yet to be agreed upon, and taking such a path would require more upfront work to support multiple SQL dialects. The other approach - a dbt package - has a known standard for integrating with dbt, a prototype could be built rather quickly, and it could leverage the multi-dialect support that comes for free in dbt. For those reasons, embedding the aql parser into a dbt package was the preferred choice for this iteration of the project.

Warehouse Compatibility

This project has been tested and is compatible with the following data warehouses:

Other warehouses (e.g. Redshift) may work but have not yet been tested.

Project Configuration

In order to use the package, a variable called dbt_activity_schema needs to be added to the project's dbt_project.yml. It should take the following format:

vars:
  dbt_activity_schema:
    column_configuration:
      included_optional_columns:
        - revenue_impact
        - link
      column_aliases:
        feature_json: attributes
    streams:
      stream_1:
        customer_id_alias: entity_uuid
        anonymous_customer_id_alias: anonymous_entity_uuid
        model_prefix: prefix_1__
      stream_2:
        customer_id_alias: entity_uuid
        model_prefix: prefix_2__

Project-Level Column Configurations

Some configurations are asserted at a project level and apply to all Activity Schema pipelines in a project. They are all related to columns included in activity and stream models, and should be nested in a second-level variable called column_configuration. They are:

included_optional_columns

Description: This variable takes a list of 0 or more columns defined as optional by the Activity Schema v2 spec. Valid options include:

Required: No

column_aliases

Description: This variable takes 0 or more nested variables, where each variable key corresponds to the standard column names as specified by the Activity Schema v2 spec:

Required: No

Stream-Level Configurations

Some configurations are unique to each Activity Schema instance.

streams

Description: This variable acts as a registry of each Activity Schema data pipeline in the project. It should contain nested variables, where each variable name corresponds to the model name of each Activity Stream model in the project. In the example above, the project should contain two Activity Stream models: stream_1 and stream_2.

Required: Yes

Nested under each registered stream are the following variables:

customer_id_alias

Description: This variable takes a string that represents the alias that will be used in lieu of the standard customer column in the Activity Schema spec, which represents the canonical ID of the entity represented by the stream.

Required: Yes

skip_stream

Description: This variable accepts a boolean true or false. When true, the downstream affects are that the build_stream macro renders an empty table, and the dataset macro queries individual activity tables directly instead of querying the stream.

Required: No (defaults to false)

anonymous_customer_id_alias

Description: This variable takes a string that represents the alias that will be used in lieu of the standard anonymous_customer_id column in the Activity Schema spec, which represents the anonymous ID of the entity represented by the stream.

Note: This is an optional column according to the Activity Schema v2 Spec. If this variable is excluded, then the Activity and Activity Stream models for the corresponding Activity Schema will exclude anonymous_customer_id as a column.

Required: No

model_prefix

Description: This variable takes a string that represents the prefix used for the names of all models associated with the corresponding Activity Schema, as it is a common practice to duplicate an activity across multiple streams, and to distinguish with a prefix in the model name itself. This variable is used for aliasing purposes downstream to improve readability of columns generated in the dataset.

Required: No

Activities

Each Activity Schema should have 1 dbt model per Activity. The model name should be the name of the Activity, and is the string that will be populated in the standard activity column. It is recommended to assert a naming convention standard for all activities for sake of consistency.

Configuration

The following variables should be added to the model's config, either in the model's yml definition or in the config block within the model itself.

An example configuration in the model config block looks like the following:

{{
    config(
        materialized='table',
        stream='stream_1',
        data_types={
            'feature_1': type_int(),
            'feature_2': type_string(),
            'feature_3': type_timestamp(),
        }
    )
}}

An example configuration in the model yml definition looks like the following:

version: 2

models:

  - name: prefix_1__activity_name
    config:
      stream: stream_1
      data_types:
        feature_1: int
        feature_2: string
        feature_3: timestamp

Note: It is highly recommended to use the first option, as it better supports with cross-database functionality.

Additional Configuration - Skipping Activity Stream

If the Activity Schema being built is configured to skip the activity stream, then each activity table should be appropriately clustered/partitioned. A convenience macro called cluster_keys is provided by this package to simplify this process. Example implementations of this configuration for supported warehouses are provided below - keep in mind that the previously referenced config parameters (stream, data_types) are still required:

-- Snowflake
{{
    config(
        cluster_by=dbt_activity_schema.cluster_keys()
    )
}}

-- Bigquery
{{
    config(
        partition_by=dbt_activity_schema.cluster_keys('stream_name').partition_by,
        cluster_by=dbt_activity_schema.cluster_keys('stream_name').cluster_by
    )
}}

-- Redshift
{{
    config(
        sort=dbt_activity_schema.cluster_keys().sort,
        dist=dbt_activity_schema.cluster_keys().dist
    )
}}

-- DuckDB
{{
    config(
        options={"partition_by": dbt_activity_schema.cluster_keys()}
    )
}}

Note: The cluster_keys macro requires the name of the activity's associated stream to be passed as an input argument only when using Bigquery.

stream

Description: The model name of the Activity Stream model in the corresponding Activity Schema pipeline. Should correspond with the name of one of the streams registered in the dbt_project.yml

Required: Yes

data_types

Description: A registry of Activity-specific columns. Expects a dictionary where each key is the alias of the column and each value is that column's data type. This information is used to automatically build the JSON object for the feature_json column in the Activity models, and to properly typecast columns when extracting them from the feature_json column when building datasets. If an Activity has no custom columns, this parameter should be excluded from the configuration.

Required: No

Code

Each Activity model should look similar to the following:

with base as (
    select
        entity_id as {{dbt_activity_schema.customer_column('stream_1')}},
        created_at as {{dbt_activity_schema.schema_columns().ts}},
        feature_1 as feature_1_alias,
        feature_2 as feature_2_alias,
        ...
    from {{ ref('my_dbt_table') }}
)
{{ dbt_activity_schema.build_activity(
    cte='base'
)}}

The build_activity macro is a convenience function that will take the data from the CTE that's passed as an argument and shape it into the standardized schema that the user has configured for the Activity model's corresponding Activity Schema. The query itself can have as much or as little business logic as necessary to define the Activity. The only expectation is that the CTE that is passed to the macro contains the following:

All other columns will be added automatically and aliased as configured in the project to the final select statement that is generated by the build_activity macro.

Generating docs and basic tests

Since each activity table has a similiar schema, it is possible to generate the documentation and add basic tests automatically. The generate_activity_yml macro takes a list of activity names as input and prints out the yaml documentation for these. You can use the output as a basis for your documentation and test setup.

Usage

In your shell use dbt run-operation to run the macro. Replace the activities list with your specific activity names as needed.

dbt run-operation generate_activity_yml --args '{activities: [customer__visited_page, customer__bought_something]}'

Copy the output to a yaml file. E.g. activites_models.yml

Input Expectations

The macro requires a single argument:

activities

Description: A list of activity names (strings). Each name in this list will generate a separate model entry in the YAML file.

Output

For each activity the macro will generate a model entry in yaml format containing columns, data types, the keys of the feature_json object as well as basic tests for uniqueness and the existence of null values.

Streams

Each Activity Schema should have exactly 1 stream model. The model should be the name of the stream that is registered in the streams variable in dbt_project.yml.

Note: This model must exist with activity model dependencies defined appropriately, regardless of how the skip_stream parameter is configured for the stream. This requirement is due to the way dbt determines model dependencies when parsing a project.

Configuration

In order to maximize computational performance when querying a stream in downstream models, it is recommended to use the clustering/partition keys as recommended by the spec. A convenience macro has been provided with adapter-specific implementations to make this easy for users. To use, simply pass the dbt_activity_schema.cluster_keys() to the appropriate argument in the model config block.

Example implementations of this configuration for supported warehouses are provided below:

-- Snowflake
{{
    config(
        cluster_by=dbt_activity_schema.cluster_keys()
    )
}}

-- Bigquery
{{
    config(
        partition_by=dbt_activity_schema.cluster_keys().partition_by,
        cluster_by=dbt_activity_schema.cluster_keys().cluster_by
    )
}}

-- Redshift
{{
    config(
        sort=dbt_activity_schema.cluster_keys().sort,
        dist=dbt_activity_schema.cluster_keys().dist
    )
}}

-- DuckDB
{{
    config(
        options={"partition_by": dbt_activity_schema.cluster_keys()}
    )
}}



Code

Each Activity model should look similar to the following:

{{ dbt_activity_schema.build_stream(
    activity_list=[
        ref('prefix_1__activity_1'),
        ref('prefix_1__activity_2')
    ]
) }}

The build_stream macro is a convenience function that takes as input a list of ref'd activity models and creates a unioned representation of them. If this model's materialization is specified as incremental, then for each activity, it will only insert rows with a ts value greater than the maximum ts value in the stream table for that activity.

Note: When skip_stream is configured as true for the stream, this macro produces an empty table.

Creating Datasets

The Activity Stream is the entrypoint for creating any denormalized dataset needed for analysis or reporting. However, as discussed above, writing bespoke SQL to query it in the standard pattern is tedious and error-prone, and using a macro to auto-generate the SQL is brittle and a subpar development experience.

To solve these shortcomings, this project introduces two solutions - a dbt macro, and an experimental DSL called aql (Activity Query Language) - an interface that's easier to learn and more SQL-esque than a macro but substantially more concise than bespoke SQL.

As a refresher, Activity Stream queries require the following inputs to be defined:

All of these inputs are still needed, and they can be viewed in the following examples - via the query_stream macro and via an aql query:

Creating Datasets Option 1: The query_stream Macro

For a macro-based approach to producing datasets, use the following syntax:

{{ dbt_activity_schema.query_stream(
    stream='stream_1',
    primary_activity=dbt_activity_schema.primary_activity(
        activity='activity_1',
        columns=[
            dbt_activity_schema.dc(
                column_name='ts',
                alias='activity_1_at',
            ),
            dbt_activity_schema.dc(
                column_name='feature_2'
            ),
            dbt_activity_schema.dc(
                column_name='feature_3',
            ),
        ],
        filters=[] -- optional (empty list is default)
    ),
    joined_activities=[
        dbt_activity_schema.appended_activity(
            activity='activity_2',
            relationship_selector='first', -- only needed for appended activities
            join_condition='ever',
            columns=[
                dbt_activity_schema.dc(
                    column_name='ts',
                    alias='first_ever_activity_2_at'
                )
            ]
        ),
        dbt_activity_schema.aggregated_activity(
            activity='activity_3',
            join_condition='after',
            columns=[
                dbt_activity_schema.dc(
                    column_name='activity_id',
                    aggfunc='count',
                    alias='count_activity_3_after'
                ),
                dbt_activity_schema.dc(
                    column_name='feature_x',
                    aggfunc='sum',
                    alias='sum_feature_x_after'
                )
            ],
            filter_columns=[], -- optional (empty list is default)
            extra_joins=[], -- optional (empty list is default)
        )
    ],
    included_columns=[] -- optional (empty list is default)
)}}

It's long, verbose, and not very readable, but this macro will produce a full sql query and return a dataset. Relevant macros and associated inputs are as follows:

Macro: query_stream

Description

Generate a dataset from an Activity Stream via a macro interface.

Args

Macro: primary_activity

A wrapper for defining the primary activity to use when building a dataset.

Args

Macro: appended_activity

A wrapper for defining activities to append-join to the primary activity use when building a dataset.

Args

Macro: aggregated_activity

A wrapper for defining activities to aggregate-join to the primary activity use when building a dataset.

Args

Macro: dc (dataset column)

A wrapper for defining columns to choose in the primary, appended, and aggregated activities in the dataset.

Args

Creating Datasets pt. 2: Querying The Activity Stream with aql

Under the hood, this package will parse the aql query string into a json object, then use the object parameters to render the appropriate SQL statement:

using stream_1
select all activity_1 (
    ts as activity_1_at,
    feature_1,
    feature_2
)
append first ever activity_2 (
    ts as first_ever_activity_2_at
)
aggregate after activity_3 (
    count(activity_id) as count_activity_3_after
    sum(feature_x) as sum_feature_x_after
)

The above statement will produce an identical dataset as the prior macro-based example in a much more concise and readable format.

To use in a dbt model, assign the aql query to a variable, and pass it to the dataset macro like so:

{% set aql %}
using stream_1
select all activity_1 (
    ts as activity_1_at,
    feature_1,
    feature_2
)
append first ever activity_2 (
    ts as first_ever_activity_2_at
)
aggregate after activity_3 (
    count(activity_id) as count_activity_3_after
    sum(feature_x) as sum_feature_x_after
)
{% endset %}

{{ dbt_activity_schema.dataset(aql) }}

Syntax details are provided below.

Activity Stream

In order to specify which Activity Stream to use, each aql query expects to begin with the following syntax:

using <stream_name>

Syntax and Parameters

using
The verb that specifies which Activity Stream to use when building the dataset. This verb must be the opening token of every aql query, or an error will be thrown.

<stream_name>
The model name of the stream in the dbt project. Under the hood, this stream will be ref'd in the SQL code derived by the package so that model lineage is maintained.

Primary Activity

In order to specify which Activity should be designated as primary, the following syntax should be used:

select <occurrence_selector> <activity_name> (
    <column>[ as <alias>],
    <column>[ as <alias>]
)

Syntax and Parameters

select
The verb that specifies which Activity will be used as the Primary Activity when building the dataset. The word that immediately follows the Activity Stream clause in the aql query must be select, or an error will be thrown.

_<occurrence_selector>_
The second word in the clause specifies which occurrence(s) of the Activity to include in the dataset. Valid options are:

<activity_name>
The last word in the clause should represent the name of an activity that exists in the Activity Stream being queried. It can include or exclude the prefix defined in the project configuration, but it should be passed as a single word without whitespaces.

<column>[ as <alias>]
For each Activity, the columns to include must be explicitly defined within parentheses. Each <column> value must correspond to the alias of one of the standard Activity Schema columns as configured in dbt_project.yml. Any <column> parsed that is not identified as such is assume to be one of the keys in the feature_json column for the Activity, and will be extracted accordingly.

Each included column can optionally be aliased to a whitespace-free friendly name as <alias>, where <alias> is the name of the column that will be applied to the dataset. Each defined alias must be preceded by as. Aliasing is optional - if no alias is explicitly defined, an automated alias will be applied to the column. See more in the alias_column macro.

Joined Activities

There are two methods of joining an Activity to the primary Activity in a dataset: append and aggregate. The join method determines the syntax used for the aql query.

Append

Joining an Activity via the append method effectively means to retrieve the column values of a specific occurrence of the Joined Activity.

append <occurrence_selector> <join_condition> <activity_name> (
    <column>[ as <alias>],
    <column>[ as <alias>]
)


Syntax and Parameters

append
The verb that specifies that the Activity will be joined to the Primary Activity via the append join method when building the dataset.

_<occurrence_selector>_
The next word in the clause specifies which occurrence of the Activity to include in the dataset from the set of Joined Activities. Valid options are:

Note: all is not a valid occurrence_selector for Appended Activities, because the append method for Joined Activites is designed to retrieve the specific column values of at most one occurrence of the Joined Activity. It is only a valid selector for the Primary Activity. Specifying all as the occurrence_selector for an Appended Activity will throw an error.


_<join_condition>_
The next word in the clause specifies how the Activity will be joined to the Primary. Valid options are:

<activity_name>
The next word in the clause should represent the name of an activity that exists in the Activity Stream being queried. It can include or exclude the prefix defined in the project configuration, but it should be passed as a single word without whitespaces.

<column>[ as <alias>]
For each Activity, the columns to include must be explicitly defined within parentheses. Each <column> value must correspond to the alias of one of the standard Activity Schema columns as configured in dbt_project.yml. Any <column> parsed that is not identified as such is assume to be one of the keys in the feature_json column for the Activity, and will be extracted accordingly.

Each included column can optionally be aliased to a whitespace-free friendly name as <alias>, where <alias> is the name of the column that will be applied to the dataset. Each defined alias must be preceded by as. Aliasing is optional - if no alias is explicitly defined, an automated alias will be applied to the column. See more in the alias_column macro.

Aggregate

Joining an Activity via the aggregate method effectively means to retrieve the column values of multiple occurrences of the Joined Activity, and roll them up via an aggregation function.

aggregate <join_condition> <activity_name> (
    <aggfunc>(<column>)[ as <alias>],
    <aggfunc>(<column>)[ as <alias>]
)


Syntax and Parameters

aggregate
The verb that specifies that the Activity will be joined to the Primary Activity via the aggregate join method when building the dataset.

_<join_condition>_
The next word in the clause specifies how the Activity will be joined to the Primary. Valid options are:

<activity_name>
The next word in the clause should represent the name of an activity that exists in the Activity Stream being queried. It can include or exclude the prefix defined in the project configuration, but it should be passed as a single word without whitespaces.

<aggfunc>(<column>)[ as <alias>]
For each Activity, the columns to include must be explicitly defined within parentheses.

For Activities joined by the aggregate method, every column needs to have an aggregation function specified. The following aggregation functions are supported:

Each <column> value must correspond to the alias of one of the standard Activity Schema columns as configured in dbt_project.yml. Any <column> parsed that is not identified as such is assume to be one of the keys in the feature_json column for the Activity, and will be extracted accordingly.

Each included column can optionally be aliased to a whitespace-free friendly name as <alias>, where <alias> is the name of the column that will be applied to the dataset. Each defined alias must be preceded by as. Aliasing is optional - if no alias is explicitly defined, an automated alias will be applied to the column. See more in the alias_column macro.

Advanced Usage

Extra Join Criteria for Joined Activities

Example code:

select all primary_activity (
    activity_id as primary_activity_id,
    ts as primary_activity_at,
    product_type as primary_activity_product_type
)
-- first joined activity
aggregate after activity_1 (
    count(activity_id) as count_activity_1_within_1_hour_after
    sum(feature_x) as sum_feature_x_within_1_hour_after
    -- join after primary activity as long as it was within 30-60 minutes of primary activity
    -- each criteria instantiated with a join, no "and"
    join {joined}.{ts} > dateadd('minutes', 30, {primary}.{ts})
    join {joined}.{ts} < dateadd('minutes', 60, {primary}.{ts})
)
-- second joined activity
append first after activity_2 (
    ts as product_type_activity_2_occurred_at
    -- join to primary activity as long as product_type value for activity_2 is equal to product_type value for primary_activity
    join json_extract_path_text({joined}.{feature_json}, 'product_type') = {primary}.primary_activity_product_type
)

Joined Activities can be further refined to the Primary Activity via extra join criteria. This functionality is useful for building datasets that represent subprocesses within a given Activity Schema (e.g. application-specific activities in a job_stream for recruiting data). This functionality is achieved by using the join keyword after all specified columns in the column selection block for a given Joined Activity. It is not applicable to Primary Activities; including this clause will be ignored. Each join criteria should be initialized with a join clause (no and clause used). It should be specified after the filter clause (if one exists in the statement). Each declared clause should relate a column from the Primary Activity to a column in the Joined Activity.

This feature supports f-string functionality for the following keywords:

Note: Any column included with the Primary Activity will be available to be used in the extra join criteria for a Joined Activity. It can be referenced via its alias and does not need to be referenced via f-string syntax (see syntax in second joined activity from example code above for reference).


⚠️ WARNING: Do not use {primary} for custom logic for Joined Activities using the aggregate join method and the all join clause. For performance purposes, these Joined Activities are not actually joined back to the Primary Activity when deriving their aggregated columns, so there will be no {primary} CTE to join to, and a SQL error will be thrown.


Defining Canonical Dataset Columns

Since some columns from joined activities will represent canonical definitions for certain business values or metrics (e.g. ts from append first ever created_account is the agreed-upon business definition of when an account became active), it's important to have a means to persist that definition and reference it by its namespaced value. That is done in two steps - the dataset_column custom materialization and the include join method in aql. It is implemented as follows:

dataset_column materialization

-- filename customer__total_items_purchased_after.sql

{% set aql %}
using customer_stream
aggregate after bought_something (
    sum(total_items_purchased)
)
{% endset %}

{{config(materialized='dataset_column', aql=aql)}}

{{ dbt_activity_schema.dataset_column(aql) }}

include join method

include (
    total_items_purchased_after -- same as filename of definition, can include or exclude configured stream prefix
)


Filtering Activities

Activities can effectively be sub-classed using the filter clause. This functionality is particularly useful for ease of development and exploratory analysis of specific subsets of activities, instead of having to create a duplicated version of that activity. Filters can be applied to Primary Activities or Joined Activities. They should be included in the column selection block, after the columns and before the extra join clauses (if any). Each filter clause should be instantiated with a filter key.

This feature supports f-string functionality for the following keywords:

⚠️ WARNING: Do not use {primary} for custom logic for Joined Activities using the aggregate join method and the all join clause. For performance purposes, these Joined Activities are not actually joined back to the Primary Activity when deriving their aggregated columns, so there will be no {primary} CTE to join to, and a SQL error will be thrown.


⚠️ WARNING: Columns contained in feature_json must be explicitly extracted and typecast. This package includes the convenience function json_extract to help - see example below for syntax specifics.


Example code:

select first visited_page (
    activity_id as activity_id,
    entity_uuid as customer_id,
    ts as first_visited_google_at
    -- json_extract will be rendered appropriately based on the target
    -- keys passed to json_extract should be wrapped in quotes
    filter {{dbt_activity_schema.json_extract('{feature_json}', 'referrer_url')}} = 'google.com'
)
append first ever visited_page (
    ts as first_2023_page_visit_at
    filter {ts} >= '2023-01-01'
)
aggregate all bought_something (
    count(activity_id) as total_large_purchases_after
    -- typecasting to non-string types (e.g. int) require including a nullif clause and an explicit typecast
    -- initiate each filter criteria with a filter keyword (not and)
    filter nullif({{dbt_activity_schema.json_extract('{feature_json}', 'total_sales')}}, '')::int > 100
    filter nullif({{dbt_activity_schema.json_extract('{feature_json}', 'total_items_purchased')}}, '')::int > 3
)

Adding Custom Aggregation Functions

Placeholder. Documentation coming soon.

Overriding Default Column Alias

Placeholder. Documentation coming soon.

Overriding Default Model Prefix Removal

Placeholder. Documentation coming soon.

Adding Custom Columns to Activity Schema

Placeholder. Documentation coming soon.

Overriding the listagg Aggregation Function Default Delimiter

Placeholder. Documentation coming soon.

Contributing

For developers who would like to contribute to this project, follow the instructions below to set up your development environment.

Setup

General Setup

  1. Fork the repository and clone the forked version
  2. Install poetry (official installer - version 1.6.1 is recommended as that is what is used in CI)
  3. Set the repo as the working directory
  4. Run poetry init to install dependencies
  5. Activate a virtual environment if not activated by default from previous step (poetry shell)
  6. Set the working directory to dbt-activity-schema/integration_tests and run dbt deps

Postgres Setup

  1. Install Postgres (brew install postgresql@15 or similar version for MacOS/Homebrew, or download the Postgres App)
  2. Start Postgres on port 5432 (should be default - brew services start postgresql@15 if using Homebrew)
  3. Create a database called dbt_activity_schema_integration_tests

Bigquery Setup

  1. Create a GCP account if you don't have one already
  2. Create a Bigquery project if you don't have one already
  3. Create a Bigquery service account if you don't have one already
  4. Obtain a JSON keyfile for the service account and save it at the path dbt-activity-schema/integration_tests/gcp_keyfile.json
  5. Add the environment variable GCP_KEYFILE_PATH=./gcp_keyfile.json

Testing

  1. After making contributions, add dbt models and corresponding dbt tests in the integration_tests subdirectory that validate the new functionality works as expected. Add tests to capture edge cases to the best of your ability. In general, for each test case, the following artifacts need to be created:
    • a dbt model
    • a dbt seed csv file that contains the expected output for the dbt model
    • a dbt_utils.equality test comparing the model and the csv file
  2. To validate that everything works, set integration_tests as the working directory, and run dbt build --target <target_name>, where <target_name> represents one of the targets defined in integration_tests/profiles.yml, and each target corresponds to a different database type. Valid options are:
    • duckdb (default)
    • bigquery
    • postgres
  3. Be sure to follow the database-specific setup instructions above to test appropriately