Additional Articles


Technical Article

PART 8: ESTIMATION

Data modeling is no doubt one of the most important and challenging aspects of developing, maintaining, augmenting and integrating typical enterprise systems. More than 90% of functionality of enterprise systems is centered round creating, manipulating and querying data. It therefore stands to reason that individuals managing enterprise projects should leverage on data modeling to execute their projects successfully and deliver not only capable and cost effective but also maintainable and extendable systems

2006-01-09

2,421 reads

Technical Article

THE ROLE OF DATA DICTIONARY

IT leaders from enterprises around the world have realized that acquiring and implementing commercial off-the-shelf (COTS) application packages is a smarter thing to do, than the traditional approach of developing home-grown applications. This may not be the fit-for-all solution, but it makes sense from a cost, time to market and resource standpoint. Selection and acquisition of the software package is just the first step. How the software package brings value through successful implementation, is critical to the business. Internal groups as well as industry analysts will closely watch COTS package implementation in the organization.

2006-01-05

2,981 reads

Technical Article

Using ITWIZ.EXE to Automate Index Analysis

This process uses ITWIZ.EXE the command line version of the Index Tuning Wizard. SQL trace file information is stored in a table on production then transferred to a non-production server. ITWIZ is then run against the transferred trace data resulting in a UNICODE SQL script. This script has index creation and deletion information as well as script parameters in a remarks section.

2005-12-29

1,973 reads

Technical Article

SQL Server 2005 Integration Services, Part 2: Developing Custom Compon

The first phase of Project REAL included the migration of existing SQL Server 2000 DTS packages to SQL Server 2005 Integration Services (SSIS) for a large electronic retailer. One source of information for the data warehouse was from TLog files, which contained data from point-of-sale cash registers in stores. TLog files store data in packed decimal format. To extract and transform this data into a format suitable for loading into the relational data warehouse tables, it was necessary to unpack and parse the files. To achieve this, custom pipeline components were implemented to read and transform the data by extending the SSIS object model. This paper discusses what we learned while implementing the pipeline components for extracting and parsing data from the TLog parser

2005-12-27

2,067 reads

Blogs

Shifting Mindsets: Why FinOps is Essential for Cloud Efficiency

By

As a DevOps practitioner, I’ve always focused on performance, scalability, and automation. But as...

March 2026 SQL Server Security Updates

By

On Patch Tuesday, in addition to OS and Office security patches, Microsoft also released...

How Fabric Mirroring Transformed with SQL Server 2025

By

When mirroring was first released for Azure SQL Database, it used Change Data Capture...

Read the latest Blogs

Forums

Samsung Galaxy A37 Price in Pakistan – Full Details and Specs

By williamscott

The Samsung Galaxy A37 is gaining attention among smartphone users who are interested in...

Multiple Deployment Processes

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Multiple Deployment Processes

How to Use sqlpackage to Detect Schema Drift Between Azure SQL Databases

By Kunal Rathi

Comments posted to this topic are about the item How to Use sqlpackage to...

Visit the forum

Question of the Day

Upgrading Admin Queries

I have a query from a former DBA that we run on SQL Server 2025 to check on database metadata. This query references sys.sysaltfiles. I want to refactor this code to be more modern. Which DMV should I reference instead?  

See possible answers