target-snowflake
Target for Snowflake.
Built with the Meltano Singer SDK.
about
stream-maps
schema-flattening
target-schema
Setting | Required | Default | Description |
---|---|---|---|
user | True | None | The login name for your Snowflake user. |
password | False | None | The password for your Snowflake user. |
private_key | False | None | The private key contents. For KeyPair authentication either private_key or private_key_path must be provided. |
private_key_path | False | None | Path to file containing private key. For KeyPair authentication either private_key or private_key_path must be provided. |
private_key_passphrase | False | None | Passphrase to decrypt private key if encrypted. |
account | True | None | Your account identifier. See Account Identifiers. |
database | True | None | The initial database for the Snowflake session. |
schema | False | None | The initial schema for the Snowflake session. |
warehouse | False | None | The initial warehouse for the session. |
role | False | None | The initial role for the session. |
add_record_metadata | False | 1 | Whether to add metadata columns. |
clean_up_batch_files | False | 1 | Whether to remove batch files after processing. |
default_target_schema | False | None | The default target database schema name to use for all streams. |
hard_delete | False | 0 | Hard delete records. |
load_method | False | TargetLoadMethods.APPEND_ONLY | The method to use when loading data into the destination. append-only will always write all input records whether that records already exists or not. upsert will update existing records and insert new records. overwrite will delete all existing records and insert all input records. |
batch_size_rows | False | None | Maximum number of rows in each batch. |
validate_records | False | 1 | Whether to validate the schema of the incoming streams. |
stream_maps | False | None | Config object for stream maps capability. For more information check out Stream Maps. |
stream_map_config | False | None | User-defined config values to be used within map expressions. |
faker_config | False | None | Config for the Faker instance variable fake used within map expressions. Only applicable if the plugin specifies faker as an addtional dependency (through the singer-sdk faker extra or directly). |
faker_config.seed | False | None | Value to seed the Faker generator for deterministic output: https://faker.readthedocs.io/en/master/#seeding-the-generator |
faker_config.locale | False | None | One or more LCID locale strings to produce localized output for: https://faker.readthedocs.io/en/master/#localization |
flattening_enabled | False | None | 'True' to enable schema flattening and automatically expand nested properties. |
flattening_max_depth | False | None | The max depth to flatten schemas. |
use_browser_authentication | False | False | If authentication should be done using SSO (via external browser). See See SSO browser authentication. |
A full list of supported settings and capabilities is available by running: target-snowflake --about
This target has an interactive feature that will help you get a Snowflake account initialized with everything needed to get started loading data.
The CLI will ask you to provide information about the new user/role/etc. you want to create but it will also need SYSADMIN credentials to execute the queries. You should prepare the following inputs:
Run the following command to get started with the interactive CLI. Note - the CLI will print the SQL queries it is planning to run and confirm with you before it makes any changes.
poetry run target-snowflake --initialize
# Alternatively using Meltano CLI
meltano invoke target-snowflake --initialize
The CLI also has a "dry run" mode that will print the queries without executing them.
Check out the demo of this on YouTube.
This Singer target will automatically import any environment variables within the working directory's
.env
if the --config=ENV
is provided, such that config values will be considered if a matching
environment variable is set either in the terminal context or in the .env
file.
You can easily run target-snowflake
by itself or in a pipeline using Meltano.
target-snowflake --version
target-snowflake --help
# Test using the "Carbon Intensity" sample:
tap-carbon-intensity | target-snowflake --config /path/to/target-snowflake-config.json
pipx install poetry
poetry install
Create tests within the target_snowflake/tests
subfolder and
then run:
poetry run pytest
You can also test the target-snowflake
CLI interface directly using poetry run
:
poetry run target-snowflake --help
Note: This target will work in any Singer environment and does not require Meltano. Examples here are for convenience and to streamline end-to-end orchestration scenarios.
Your project comes with a custom meltano.yml
project file already created.
Next, install Meltano (if you haven't already) and any needed plugins:
# Install meltano
pipx install meltano
# Initialize meltano within this directory
cd target-snowflake
meltano install
Now you can test and orchestrate using Meltano:
# Test invocation:
meltano invoke target-snowflake --version
# OR run a test `elt` pipeline with the Carbon Intensity sample tap:
meltano run tap-carbon-intensity target-snowflake
See the dev guide for more instructions on how to use the Meltano SDK to develop your own Singer taps and targets.