Additional Articles


External Article

Using the Data Profiling SQL Server Integration Services SSIS task

Have you ever had to write a bunch of TSQL to do some data analysis on the table data in your database. If you have you'll know that this can become a fairly time consuming and tedious task? SQL Server 2012 Integration Services has a feature called the Data Profiling task that can help you perform this analysis much easier and faster (this feature is also available in SQL Server 2008). This task can help you find potential issues with your existing data as well as help you become more familiar with the data in a database that you have just started managing.

2013-02-04

3,103 reads

External Article

Providing SQL Agent Job Log Data for Developers

As a production Database Administrator, I do not want to give Developers direct access to the SQL Server Agent job log especially for servers in the DMZ. Another problem we have is that when there is a lot of log data for a job, the default job log doesn't contain the full log detail and that makes it harder to troubleshoot. Most of all, we are trying not to use a different code set for deployment based on the environment. Basically we want to use the same methods to deploy our jobs to Development, Test and Production.

2013-01-29

1,816 reads

Technical Article

Beating Backup Corruption

The most critical task for all DBAs is to have a Backup and Recovery strategy that ensures, every day, that in the event of a disaster they can restore and recover any database, within acceptable limits for data loss and downtime. Even with all the required backups in place, it's easy to miss subtle failings in the overall plan that can, and eventually will, defeat your recovery plans.

You rated this post out of 5. Change rating

2013-01-28

2,889 reads

External Article

We Loaded 1TB in 30 Minutes with SSIS, and So Can You

In February 2008, Microsoft announced a record-breaking data load using Microsoft® SQL Server® Integration Services (SSIS): 1 TB of data in less than 30 minutes. That data load, using SQL Server Integration Services, was 30% faster than the previous best time using a commercial ETL tool. This paper outlines what it took: the software, hardware, and configuration used. We will describe what we did to achieve that result, and offer suggestions for how to relate these techniques to typical scenarios. Even for customers who don't have needs quite like this benchmark, such efforts can teach a lot about getting optimal performance.

2013-01-25 (first published: )

9,371 reads

Blogs

Shifting Mindsets: Why FinOps is Essential for Cloud Efficiency

By

As a DevOps practitioner, I’ve always focused on performance, scalability, and automation. But as...

March 2026 SQL Server Security Updates

By

On Patch Tuesday, in addition to OS and Office security patches, Microsoft also released...

How Fabric Mirroring Transformed with SQL Server 2025

By

When mirroring was first released for Azure SQL Database, it used Change Data Capture...

Read the latest Blogs

Forums

Multiple Deployment Processes

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Multiple Deployment Processes

How to Use sqlpackage to Detect Schema Drift Between Azure SQL Databases

By Kunal Rathi

Comments posted to this topic are about the item How to Use sqlpackage to...

Upgrading Admin Queries

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Upgrading Admin Queries

Visit the forum

Question of the Day

Upgrading Admin Queries

I have a query from a former DBA that we run on SQL Server 2025 to check on database metadata. This query references sys.sysaltfiles. I want to refactor this code to be more modern. Which DMV should I reference instead?  

See possible answers