Additional Articles


External Article

Using the Data Profiling SQL Server Integration Services SSIS task

Have you ever had to write a bunch of TSQL to do some data analysis on the table data in your database. If you have you'll know that this can become a fairly time consuming and tedious task? SQL Server 2012 Integration Services has a feature called the Data Profiling task that can help you perform this analysis much easier and faster (this feature is also available in SQL Server 2008). This task can help you find potential issues with your existing data as well as help you become more familiar with the data in a database that you have just started managing.

2013-02-04

3,103 reads

External Article

Providing SQL Agent Job Log Data for Developers

As a production Database Administrator, I do not want to give Developers direct access to the SQL Server Agent job log especially for servers in the DMZ. Another problem we have is that when there is a lot of log data for a job, the default job log doesn't contain the full log detail and that makes it harder to troubleshoot. Most of all, we are trying not to use a different code set for deployment based on the environment. Basically we want to use the same methods to deploy our jobs to Development, Test and Production.

2013-01-29

1,816 reads

Technical Article

Beating Backup Corruption

The most critical task for all DBAs is to have a Backup and Recovery strategy that ensures, every day, that in the event of a disaster they can restore and recover any database, within acceptable limits for data loss and downtime. Even with all the required backups in place, it's easy to miss subtle failings in the overall plan that can, and eventually will, defeat your recovery plans.

You rated this post out of 5. Change rating

2013-01-28

2,889 reads

External Article

We Loaded 1TB in 30 Minutes with SSIS, and So Can You

In February 2008, Microsoft announced a record-breaking data load using Microsoft® SQL Server® Integration Services (SSIS): 1 TB of data in less than 30 minutes. That data load, using SQL Server Integration Services, was 30% faster than the previous best time using a commercial ETL tool. This paper outlines what it took: the software, hardware, and configuration used. We will describe what we did to achieve that result, and offer suggestions for how to relate these techniques to typical scenarios. Even for customers who don't have needs quite like this benchmark, such efforts can teach a lot about getting optimal performance.

2013-01-25 (first published: )

9,371 reads

Blogs

Ramblings about data communities and your contributions, no excuses

By

I have been active in the data community throughout my career. I have met...

SQL Server Journey till 2025 (brief)

By

Quick Summary for Microsoft SQL Server till 2025, I am fortunate to be part...

Building Power BI Reports: Desktop vs Fabric

By

Why this comparison feels confusing If you’re a Power BI report author who’s just...

Read the latest Blogs

Forums

Simulating Mercury’s Orbital Motion Using Pure T-SQL (NASA 2025 Dataset)

By NKTgLaw

SQL Server is typically viewed as a transactional or analytical database engine. However, it...

Azure SQL Database | CREATE EVENT SESSION [Blocking_Capture] ON DATABASE

By DanielP

Hello, Is there a way in Azure SQL Database to change the 'Blocking Process...

Having a Little Fun at SQL Server Central

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Having a Little Fun at...

Visit the forum

Question of the Day

String Similarity I

On SQL Server 2025, when I run this, what is returned?

SELECT EDIT_DISTANCE_SIMILARITY('SQL Server', 'MySQL')

See possible answers