Samuel Nogueira Farrus
02/11/2025, 12:14 PMtap-postgres target-postgres
EL to read and store a table with 5,500,000+ rows and 45+ columns (and unfortunately, I need them all). With use_copy: true
, the data ingestion takes an hour to complete, while in other solutions it would take less than 10 minutes (as expected of a COPY
statement). Any ideas on how can I optimise this??Samuel Nogueira Farrus
02/11/2025, 7:35 PMtap-postgres
is the real issue: I added a CSV in between (so tap-postgres target_csv
and tap-csv target-postgres
). The first takes almost an hour, while the last took 6-7 minutes. So any ideas on how to optimise (speed-up) the data input in the tap-postgres
?? Tried the configurations, but found nothing helpful in this matterEdgar Ramírez (Arch.dev)
02/11/2025, 11:03 PMSamuel Nogueira Farrus
02/13/2025, 8:03 PMSamuel Nogueira Farrus
02/25/2025, 1:47 PMtarget-postgres
batch_processing_time is an average of 20 s to each batch, and it sums to ~6 minutes of total COPY
operation time, which is acceptable.
The tap-postgres
average extraction time interval between records is ~0.001 s, what leads (roughly) to a total of 90 minutes of data reading.Edgar Ramírez (Arch.dev)
02/25/2025, 3:57 PMSamuel Nogueira Farrus
02/25/2025, 4:30 PMEdgar Ramírez (Arch.dev)
02/25/2025, 9:41 PM