Hello: I'm looking to load a non-trivial amount of data into dataverse (~10-15 million records for 2 entities and a several that are between 500k and 1 million) using Synapse Pipelines. We're looking to plan for the production move and need to get a rough estimate of how long it will take (and ideally upgrade as necessary to keep this time as small as possible). I have read through the API documentation but I was wondering if there's a way to determine what the load times will be. For instance, if 1million records takes 8 hours, I don't believe it's linear in calculating that 12 million would be 12*8 . I believe the Sandbox instances are a good bit slower in general than production ones but I'm not sure how to quantify it.
1-Is there something we can add in terms of licensing that will speed up performance on loads , assuming that we're sure the bottlenecks are on the Dataverse side?
2-Is it correct that there's a fairly signficant difference between Sandbox and Prod instances? I am pretty sure there is, but if so, is there any way to approximately quantify the differences?
3-Anyone done any loads of this size and have any general thoughts or suggestions?
Thank you.
Hi @Bill_Ryan33149,
I dont recall if there is published information on how resources are allocated to servers exactly, but I do believe the environment is put on an appropriate VM tier based on a number of factors like number of licensed users in the environment, number of transactions, database size, and other metrics. We have no control over this.
Things that will improve performance include:
Hope this helps.
---
Please click Accept as Solution if my post answered your question. This will help others find solutions to similar questions. If you like my post and/or find it helpful, please consider giving it a Thumbs Up.
stampcoin
17
mmbr1606
15
Super User 2025 Season 1
ankit_singhal
11
Super User 2025 Season 1