and in that entire time it didn’t upload any recor...
# troubleshooting
c
and in that entire time it didn’t upload any records to redshift. it’s like it’s waiting for the entire data set to be downloaded before starting any uploading?
x
exactly as what I saw, it doesn’t load anything to my postgres 🥲 I got the same feeling
t
Are you seeing any records output on the command line?
d
@chris_kings-lynne You may want to tweak this setting: https://meltano.com/plugins/loaders/redshift.html#batch-size-rows
c
This is a tiny sample of a 13 hour run that failed when it went OOM. During this period, exactly zero rows were written to Redshift. It also seems to spend like an hour to extact 100 records
@douwe_maan no point really tweaking that as if the run doesn’t finish the state won’t be written and that data in redshift doesn’t help us right
t
I’ve tested tap-zendesk locally before and it seemed to work with a smaller set of data - I think I was extracting just a month’s worth of tickets. It sounds like it’s something with the tap just not sending records.
c
I’ve still not managed to get a 12 month refresh to complete
latest failure after 4 hours is:
Untitled
I think at this point (a) the zendesk api is insanely slow/rate limited
and (b) the tap is not coded defensively against running for a very long time
With that run I was using redshift flush all streams at 500 records, and I am getting rows in redshift
d
@chris_kings-lynne So target-redshift does not output state until the very end? Or tap-zendesk itself doesn't? We should address that so that even an interrupted run can be recovered
c
It does seem to update state now after each batch upload, so we’re good. the problem was that in a 14 hour run, tap zendesk never got to the 100k default rows for redshift