I have a dataflow that takes data from sharepoint in csv form, it then filters the data and writes it into the dataverse table. The nature of the system the csv data is extracted from means we get an ever growing file that resets at the end of the year, so we don't want to load all of the data in the file everytime we get it, just the new entries that have been added to it need loaded into the dataverse.
It took me a while to figure out but I now have a way of identifying what's new and old in the file, I have a key setup on the dataverse table that uniquely identifies each row, the query in the dataverse table is setup to merge into the dataverse on this key column using a left anti join which should "bring in only rows from the left table that don't have any matching rows from the right table" why would the dataflow be doing anything to a row in the table that it knows based on this key not to update?
Any ideas?