Additional Articles


External Article

Using the Data Profiling SQL Server Integration Services SSIS task

Have you ever had to write a bunch of TSQL to do some data analysis on the table data in your database. If you have you'll know that this can become a fairly time consuming and tedious task? SQL Server 2012 Integration Services has a feature called the Data Profiling task that can help you perform this analysis much easier and faster (this feature is also available in SQL Server 2008). This task can help you find potential issues with your existing data as well as help you become more familiar with the data in a database that you have just started managing.

2013-02-04

3,103 reads

External Article

Providing SQL Agent Job Log Data for Developers

As a production Database Administrator, I do not want to give Developers direct access to the SQL Server Agent job log especially for servers in the DMZ. Another problem we have is that when there is a lot of log data for a job, the default job log doesn't contain the full log detail and that makes it harder to troubleshoot. Most of all, we are trying not to use a different code set for deployment based on the environment. Basically we want to use the same methods to deploy our jobs to Development, Test and Production.

2013-01-29

1,816 reads

Technical Article

Beating Backup Corruption

The most critical task for all DBAs is to have a Backup and Recovery strategy that ensures, every day, that in the event of a disaster they can restore and recover any database, within acceptable limits for data loss and downtime. Even with all the required backups in place, it's easy to miss subtle failings in the overall plan that can, and eventually will, defeat your recovery plans.

You rated this post out of 5. Change rating

2013-01-28

2,889 reads

External Article

We Loaded 1TB in 30 Minutes with SSIS, and So Can You

In February 2008, Microsoft announced a record-breaking data load using Microsoft® SQL Server® Integration Services (SSIS): 1 TB of data in less than 30 minutes. That data load, using SQL Server Integration Services, was 30% faster than the previous best time using a commercial ETL tool. This paper outlines what it took: the software, hardware, and configuration used. We will describe what we did to achieve that result, and offer suggestions for how to relate these techniques to typical scenarios. Even for customers who don't have needs quite like this benchmark, such efforts can teach a lot about getting optimal performance.

2013-01-25 (first published: )

9,371 reads

External Article

Hypothetical Indexes on SQL Server

Sometimes, you have to test out alternative indexing strategies, but the task of creating the candidate indexes would just take too long. Is there another way? Well, yes, why not use the same method the DTA (Database Tuning Advisor) uses, and take the tedium out of the job.

2013-01-22

3,343 reads

Blogs

Introducing the SQL Server on Kubernetes Operator

By

Are you considering replatforming your SQL Server workload due to recent vendor changes, but...

Advice I Like: Rewards from Work

By

The greatest rewards come from working on something that nobody has words for. If...

Overcoming Challenges: Navigating Common Pitfalls in FinOps Adoption

By

Working in DevOps, I’ve seen FinOps do amazing things for cloud cost control, but...

Read the latest Blogs

Forums

The day-to-day pressures of a DBA team, and how we can work smarter with automation and AI

By Terry Jago

Comments posted to this topic are about the item The day-to-day pressures of a...

Analysis Services Model w/ Direct Query and (Default Veritpaq)

By Archivist

Analysis Services (either the integrated workspace in Power BI or on a SQL Server)...

Don't Panic

By Grant Fritchey

Comments posted to this topic are about the item Don't Panic

Visit the forum

Question of the Day

Identities and Sequences IV

When thinking about the identity property and sequence objects, which of these can be used with numeric and decimal data types?

See possible answers