February 28, 2011 at 5:15 pm
Hi All,
I am uploading data of one table from one server to another server. The table contains 62 million rows for one month. For better optimization I designed the package using BCP out and BulkInsert task. It is taking 45 mins to transfer the data....
Could any one please suggest to speed up the package. Can I use any other task to reduce the time.
Or need to modify anything in BulkInsert task like Batchsize,options etc....?
If I use Data Flow task it taking more than one hour. So, I have choosen BCP and bulk insert task.
Thanks in advance.
your help is appreciated.
Divya
March 1, 2011 at 3:44 am
The following article can have some great suggestions for you:
Not all best practices can be applied to your situation, but some of them ought to be useful.
Regarding the INSERTS in the destination:
* make sure that there are no constraints (foreign key or check)
* drop indexes before the insert and recreate them after the package has done
* set the recovery model to bulk-logged or simple
Need an answer? No, you need a question
My blog at https://sqlkover.com.
MCSE Business Intelligence - Microsoft Data Platform MVP
Viewing 2 posts - 1 through 2 (of 2 total)
You must be logged in to reply to this topic. Login to reply