Viewing 15 posts - 421 through 435 (of 1,124 total)
This should do:
RIGHT( FIRST_NAME, LEN( FIRST_NAME ) - PATINDEX( ' ', FIRST_NAME ) ) + LAST_NAME
March 16, 2009 at 10:03 am
I just can't recollect now, can you just cross check for the option "TEXT IDENTIFIER" in BULK INSERT?
foxjazz (3/14/2009)
However...
March 14, 2009 at 11:19 am
ROW_NUMBER() is what are looking for, Check Books Online for more info..
SELECT ROW_NUMBER() OVER( PARTITION BY Col2 ORDER BY Col2 ) AS SeqNo, Col2
FROM SomeTable
March 14, 2009 at 11:13 am
I feel very upset that I don't have any SQL right now:crazy::crying:, Sorry michael:ermm:, you have to wait until Monday if nobody else beat it to me.:D
March 14, 2009 at 11:06 am
Jeff Moden (3/14/2009)
Ramesh (3/14/2009)
See this article on Looping, Importing and Archiving files in DTS Packages using ActiveX Script
Just do 3 simple bulk inserts. It's also one of the fastest...
March 14, 2009 at 11:00 am
You just need 2 connections & 1 data flow task, within the flow task you will have source and destination components. And in the source, you select SQL command...
March 14, 2009 at 10:33 am
See this article on Looping, Importing and Archiving files in DTS Packages using ActiveX Script
March 14, 2009 at 9:55 am
Does the queries are part of a procedure/function that were created using RECOMPILE option? This may be causing the server to re-compile every time on each execution.
March 14, 2009 at 7:40 am
No, any change in database properties does not require a server restart.
March 14, 2009 at 6:41 am
Does the query returns large no. of rows? How much time does it take to execute? How about adding the execution plan as an attachment, so to check for...
March 14, 2009 at 6:39 am
Still not very clear, why don't you read this article on how to post questions to get better/quick responses[/url] and help us in helping you.
March 14, 2009 at 6:17 am
What errors do you get when connecting to server in your web application?
Can you able to connect to SQL Server using SSMS from application server?
March 14, 2009 at 6:04 am
No, the data will not be deleted. The Transform Data task just dumps data from source to destination.
March 14, 2009 at 5:56 am
Viewing 15 posts - 421 through 435 (of 1,124 total)