adam_rudd
10/01/2021, 3:26 AMpipeline1: Pulling from SalesForce and loading via target-s3-csv to an s3 bucket with account role ARN:123 `salesforce-data`
pipeline2: Pulling from Jira and loading via target-s3-csv to an s3 bucket with account role ARN:456 `salesforce-data`
If dockerizing this, i’m under the impression that we should have a single meltano
container with a bunch of workers which are used for actual processing.
TLDR: whats the recommended approach for re-using targets with different configs (writing to the same location is not feasible for us so we cannot consider this option unfortunately)boggdan_barrientos
10/01/2021, 3:56 AMinherit_from:
works for tap/targets you only have to change the configuration. Something like this.
- name: target-redshift
variant: transferwise
pip_url: pipelinewise-target-redshift
config:
primary_key_required: false
batch_size_rows: 100000
- name: target-redshift-raw
inherit_from: target-redshift
config:
primary_key_required: true
batch_size_rows: 250000
parallelism: -1
adam_rudd
10/01/2021, 5:26 AMedgar_ramirez_mondragon
10/01/2021, 3:13 PM$MELTANO_EXTRACT__LOAD_SCHEMA
(for example target-postgres). That variable is filled at runtime from the extractor definition: https://meltano.com/docs/plugins.html#load-schema-extra. So if you have an extractor for Salesforce with a namespace salesforce
, Jira with namespace jira
and a target-s3-csv that defines s3_bucket: $MELTANO_EXTRACT__LOAD_SCHEMA
, then each source will land in a separate bucket in S3