Hi, we haven’t had any issues. We considered self-hosting, but ended up going hosted via confluent, it is pricier, but we’ve done some optimization to get that down which would work in a self-hosted environment as well. We are considering transitioning over to self-hosted (we self-host meltano and looking to self-host observability tools as well).
Haven’t had any issues for a full year. We went with kafka primarily for multiple consumer route as we had s3/snowflake and other destinations to push to. Additionally, you may see a lower cost for your backend if self-hosting (i.e. your hosting cost), but depending on your destination you may see elevated compute cost. For us with the amount of data we’d be writing into snowflake as an example we need a streaming based connector which doesn’t use a virtual warehouse for inserts. I assume you would need a virtual warehouse for all snowflake connectors in meltano which makes that potentially pricier alternative especially if batch and inserting happens simultaneously. We had that issue when we considered stitch for cdc (cost on stitch side not bad, but snowflake cost was very high). Depending on your use case (i.e. your destination) this may not be as relevant.
Of course we’re connecting to a master instance so disk utilization is important, but as long as you configure things correctly (i.e. implementing heartbeats as an example) you should be confident in the streaming pipeline, while keeping your master instance online.