It seems the problem is the nvarchar(max) field when it contains too much data. I changed the field to varbinary(max) and now the synchronization of the same data is over in a fraction of the time used before.
If you are using the client/server libraries, the communication ends up being essentially a dataset that is serialized into XML (it isn't technically "serialized", but it ends up being roughly similar). So you start with the dataset which has a reputation for being bloated. Additionally the dataset will store before and after data for each row that is changed - doubling the amount of data.
Then add serialization to XML which adds more overhead. Then that payload is wrapped in web service calls - which add a little more overhead. All told, you are looking at transferring a very large amount of data across the wire.
The time lapse you are seeing is just the amount of time it is taking to transfer all that data.
I cant really gove you an answer,but what I can give you is a way to a solution, that is you have to find the anglde that you relate to or peaks your interest. A good paper is one that people get drawn into because it reaches them ln some way.As for me WW11 to me, I think of the holocaust and the effect it had on the survivors, their families and those who stood by and did nothing until it was too late.