Additional Articles


Technical Article

PART 8: ESTIMATION

Data modeling is no doubt one of the most important and challenging aspects of developing, maintaining, augmenting and integrating typical enterprise systems. More than 90% of functionality of enterprise systems is centered round creating, manipulating and querying data. It therefore stands to reason that individuals managing enterprise projects should leverage on data modeling to execute their projects successfully and deliver not only capable and cost effective but also maintainable and extendable systems

2006-01-09

2,421 reads

Technical Article

THE ROLE OF DATA DICTIONARY

IT leaders from enterprises around the world have realized that acquiring and implementing commercial off-the-shelf (COTS) application packages is a smarter thing to do, than the traditional approach of developing home-grown applications. This may not be the fit-for-all solution, but it makes sense from a cost, time to market and resource standpoint. Selection and acquisition of the software package is just the first step. How the software package brings value through successful implementation, is critical to the business. Internal groups as well as industry analysts will closely watch COTS package implementation in the organization.

2006-01-05

2,981 reads

Technical Article

Using ITWIZ.EXE to Automate Index Analysis

This process uses ITWIZ.EXE the command line version of the Index Tuning Wizard. SQL trace file information is stored in a table on production then transferred to a non-production server. ITWIZ is then run against the transferred trace data resulting in a UNICODE SQL script. This script has index creation and deletion information as well as script parameters in a remarks section.

2005-12-29

1,973 reads

Technical Article

SQL Server 2005 Integration Services, Part 2: Developing Custom Compon

The first phase of Project REAL included the migration of existing SQL Server 2000 DTS packages to SQL Server 2005 Integration Services (SSIS) for a large electronic retailer. One source of information for the data warehouse was from TLog files, which contained data from point-of-sale cash registers in stores. TLog files store data in packed decimal format. To extract and transform this data into a format suitable for loading into the relational data warehouse tables, it was necessary to unpack and parse the files. To achieve this, custom pipeline components were implemented to read and transform the data by extending the SSIS object model. This paper discusses what we learned while implementing the pipeline components for extracting and parsing data from the TLog parser

2005-12-27

2,067 reads

Blogs

“We love to debate minutiae”

By

I am guilty as charged. The quote was in reference to how people argue...

Advice I Like: Knots

By

Learn how to tie a bowline knot. Practice in the dark. With one hand....

Shifting Mindsets: Why FinOps is Essential for Cloud Efficiency

By

As a DevOps practitioner, I’ve always focused on performance, scalability, and automation. But as...

Read the latest Blogs

Forums

Windows logins for users migrated from DomainA to DomainB

By a.koopman

Hi, I have a SQL Server instance where users connect to via Windows Authentication,...

Multiple Deployment Processes

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Multiple Deployment Processes

How to Use sqlpackage to Detect Schema Drift Between Azure SQL Databases

By Kunal Rathi

Comments posted to this topic are about the item How to Use sqlpackage to...

Visit the forum

Question of the Day

Upgrading Admin Queries

I have a query from a former DBA that we run on SQL Server 2025 to check on database metadata. This query references sys.sysaltfiles. I want to refactor this code to be more modern. Which DMV should I reference instead?  

See possible answers