Hi,- I have a scenario to load a table with close to 100 million records from MSSQL server(on-premise) to Bigquery in GCP. I am using tap-mssql extractor and target-bigquery loader(default variants). Can you tell me settings/configurations which need to be set in both extractor ad loader in meltano.yml to complete the data load fast and in an efficient manner ?
I am using tap-mysql to load about 50 million rows of data and experiencing speed about 1000 rows/seconds on our private cluster.
I am using stream-maps so BATCH mode is not yet supported. @Edgar Ramírez (Arch.dev) Do you have any suggestions to speed this up?
The flow is incremental so I can live with that speed later on but of course faster is welcome.
e
Edgar Ramírez (Arch.dev)
06/13/2024, 10:45 AM
Would you be able to identify the source of the slowness. If so, we can address it directly in the tap and it'd be a quick win for all users of it.
v
Viet Vu Danh
06/14/2024, 2:16 AM
Yeah, I am diggingg in but not sure where the bottle neck might is.
Maybe I am using stream maps? Or different variant will have different performance (I need ssl auth so stucked with transferwise variant)? I will try different config/settings when I have time to test.