Hi y’all!
I am looking to do a migration from Postgres to snowflake using log-based replication. I would like to know if there are any best practices/pro tips/well known techniques for a pipeline that uses log-based replication from the start (to do an initial full dump/sync and then orchestrate incremental uploads using log-based).
Some side notes and observations…
• in staging, the tables are small enough to manage.
◦ But in production, the tables are pretty huge.
• When running the log-based pipeline in staging, it requires a few runs to get all the data migrated over.
◦ Also, within that same migration, one table is ~50k rows in size when it should be >200.
Thank you so much for the help 😁
cc: @luke_winslow