Hello: I'm looking to load a non-trivial amount of data into dataverse (~10-15 million records for 2 entities and a several that are between 500k and 1 million) using Synapse Pipelines. We're looking to plan for the production move and need to get a rough estimate of how long it will take (and ideally upgrade as necessary to keep this time as small as possible). I have read through the API documentation but I was wondering if there's a way to determine what the load times will be. For instance, if 1million records takes 8 hours, I don't believe it's linear in calculating that 12 million would be 12*8 . I believe the Sandbox instances are a good bit slower in general than production ones but I'm not sure how to quantify it.
1-Is there something we can add in terms of licensing that will speed up performance on loads , assuming that we're sure the bottlenecks are on the Dataverse side?
2-Is it correct that there's a fairly signficant difference between Sandbox and Prod instances? I am pretty sure there is, but if so, is there any way to approximately quantify the differences?
3-Anyone done any loads of this size and have any general thoughts or suggestions?