SQL Clone
SQLServerCentral is supported by Redgate
 
Log in  ::  Register  ::  Not logged in
 
 
 


The Multi Phase Data Pump


The Multi Phase Data Pump

Author
Message
Dinesh Asanka
Dinesh Asanka
Ten Centuries
Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)

Group: General Forum Members
Points: 1290 Visits: 223
Comments posted to this topic are about the content posted at http://www.sqlservercentral.com/columnists/dasanka/themultiphasedatapump.asp






My Blog: http://dineshasanka.spaces.live.com/
philcart
philcart
SSCarpal Tunnel
SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)

Group: General Forum Members
Points: 4118 Visits: 1436

Nice article, but would have liked a mention that using the Multi-phase Datapump to transform data will switch the datapump to row-by-row processing. This is something that people should be aware of and can lead to data loads taking much longer than if a normal datapump copy column task is used.

Also, you don't have to use all the phases. In a lot of my packages I use just the PumpComplete phase. The VBScript in this phase uses the DTSTransformPhaseInfo object to capture the number of rows exported to the file. This object also has an ErrorRows property that holds the number of errors encountered.



Hope this helps
Phill Carter
--------------------
Colt 45 - the original point and click interface

Australian SQL Server User Groups - My profile
Phills Philosophies
Murrumbeena Cricket Club
Dinesh Asanka
Dinesh Asanka
Ten Centuries
Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)

Group: General Forum Members
Points: 1290 Visits: 223

yes u are correct phillcart.

I just enable all the phases just to give an idea as u said it is not a compulsory thing to do.








My Blog: http://dineshasanka.spaces.live.com/
Peter Kryszak
Peter Kryszak
Ten Centuries
Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)Ten Centuries (1.3K reputation)

Group: General Forum Members
Points: 1348 Visits: 3
Our daily DTS data pumps normally run without errors, but there are a lot of messages written to the log file. If errors do occur, the messages don't help much and we have to investigate the data file in any case. It seems that we can probably reduce those messages to indicate only success or failure by using a tailored multiphase data pump.

Please comment on the differences between the default and a multiphase data pump in terms of what the default actions are. What might be involved in creating a less verbose data pump?



philcart
philcart
SSCarpal Tunnel
SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)SSCarpal Tunnel (4.1K reputation)

Group: General Forum Members
Points: 4118 Visits: 1436

As I mentioned above the main difference is if you use the Multi Phse datapump in data transformations. This switches the datapump into row-by-row processing instead of a bulk load.

If you are able to accept the performance hit you can develop a pretty sophisticated data loadng routine that traps data conversion and key violation errors. The rows in error can be redirected to an exception table for later examination.

If you're loading a lot of data then a better approach would be to bulk load the file directly into a staging table that has all the fields defined as varchar. Then you can run validation routines on the data without affecting performance.



Hope this helps
Phill Carter
--------------------
Colt 45 - the original point and click interface

Australian SQL Server User Groups - My profile
Phills Philosophies
Murrumbeena Cricket Club
Go


Permissions

You can't post new topics.
You can't post topic replies.
You can't post new polls.
You can't post replies to polls.
You can't edit your own topics.
You can't delete your own topics.
You can't edit other topics.
You can't delete other topics.
You can't edit your own posts.
You can't edit other posts.
You can't delete your own posts.
You can't delete other posts.
You can't post events.
You can't edit your own events.
You can't edit other events.
You can't delete your own events.
You can't delete other events.
You can't send private messages.
You can't send emails.
You can read topics.
You can't vote in polls.
You can't upload attachments.
You can download attachments.
You can't post HTML code.
You can't edit HTML code.
You can't post IFCode.
You can't post JavaScript.
You can post emoticons.
You can't post or upload images.

Select a forum

































































































































































SQLServerCentral


Search