xiaozhou_wang
11/29/2022, 10:53 PMmeltano run tap-postgres target-jsonl
I'm using S3 as the backend and then using aws cli to upload the jsonl into an s3 bucket.
This process all works fine initially but I run into problems when updating the select
pattern for tables in tap-postgres. After the run completes, there are no entries added to state.json for the new tables. Therefore, every run ends up with those tables getting a full replication. The state.json is getting updated - update timestamp shows this and also the log serial number for the old tables updates. The logs also do not indicate problems:
2022-11-29T22:01:41.099318Z [info ] Reading state from AWS S3
2022-11-29T22:01:41.559899Z [info ] smart_open.s3.MultipartWriter('XXXXX', 'meltano_state/prod:tap-postgres-to-target-jsonl/lock'): uploading part_num: 1, 17 bytes (total 0.000GB)
....
2022-11-29T22:02:03.196648Z [info ] Writing state to AWS S3
2022-11-29T22:03:42.028608Z [info ] smart_open.s3.MultipartWriter('XXXX', 'meltano_state/prod:tap-postgres-to-target-jsonl/lock'): uploading part_num: 1, 17 bytes (total 0.000GB)
2022-11-29T22:03:42.243697Z [info ] smart_open.s3.MultipartWriter('XXXX', 'meltano_state/prod:tap-postgres-to-target-jsonl/state.json'): uploading part_num: 1, 774 bytes (total 0.000GB)
2022-11-29T22:03:42.382744Z [info ] Incremental state has been updated at 2022-11-29 22:03:42.382636.
I know the docs recommend a postgres backend but I would rather not set up a postgres db just for this. The s3 persistence should work since I'm just running from one source!xiaozhou_wang
11/29/2022, 11:03 PMxiaozhou_wang
11/30/2022, 9:29 AMcody_hanson
11/30/2022, 4:30 PMxiaozhou_wang
11/30/2022, 5:05 PM