jon_lindstrom_bolmgren
02/06/2023, 3:23 PMDon’t just provide the “Airflow-OracleDB-Ingester” as a process, but the “Airflow-OracleDB-Ingester-ID1244” where ID1244 clearly identifies the specific run of ingestion.What would be the best way to add such metadata columns? I use a mix of different extractors together with
target-postgres
(transferwise) and Airflow for orchestration. I saw that target-postgres
has its own config setting for metadata columns, but these are a bit different from whats suggested in the article. Does Meltano offer some way to inject additional attributes during loading? 🤔taylor
02/06/2023, 3:33 PMadd_record_metadata=True
and you’ll get a bunch of useful fields added https://sdk.meltano.com/en/latest/implementation/record_metadata.html
Another way to do this would be to use stream maps to add a new column https://sdk.meltano.com/en/latest/stream_maps.html#built-in-functions cc @Sven Balnojan @aaronsteerspat_nadolny
02/06/2023, 4:09 PMadd_metadata_columns
that you can use alsojon_lindstrom_bolmgren
02/06/2023, 4:31 PMadd_record_metadata
setting for Meltano SDK based targets. Since I currently use the transferwise target-postgres
I'll experiment with its add_metadata_columns
setting. In addition to these timestamps, I liked the idea of having the Airflow run ID as metadata, but looks like this is only possible via some type of custom stream mapping.Matt Menzenski
02/06/2023, 5:40 PMThe transferwise target-postgres has a setting calledthe transferwise target-redshift plugin does too, we use itthat you can use alsoadd_metadata_columns