Additional Articles


Technical Article

PART 8: ESTIMATION

Data modeling is no doubt one of the most important and challenging aspects of developing, maintaining, augmenting and integrating typical enterprise systems. More than 90% of functionality of enterprise systems is centered round creating, manipulating and querying data. It therefore stands to reason that individuals managing enterprise projects should leverage on data modeling to execute their projects successfully and deliver not only capable and cost effective but also maintainable and extendable systems

2006-01-09

2,421 reads

Technical Article

THE ROLE OF DATA DICTIONARY

IT leaders from enterprises around the world have realized that acquiring and implementing commercial off-the-shelf (COTS) application packages is a smarter thing to do, than the traditional approach of developing home-grown applications. This may not be the fit-for-all solution, but it makes sense from a cost, time to market and resource standpoint. Selection and acquisition of the software package is just the first step. How the software package brings value through successful implementation, is critical to the business. Internal groups as well as industry analysts will closely watch COTS package implementation in the organization.

2006-01-05

2,981 reads

Technical Article

Using ITWIZ.EXE to Automate Index Analysis

This process uses ITWIZ.EXE the command line version of the Index Tuning Wizard. SQL trace file information is stored in a table on production then transferred to a non-production server. ITWIZ is then run against the transferred trace data resulting in a UNICODE SQL script. This script has index creation and deletion information as well as script parameters in a remarks section.

2005-12-29

1,973 reads

Technical Article

SQL Server 2005 Integration Services, Part 2: Developing Custom Compon

The first phase of Project REAL included the migration of existing SQL Server 2000 DTS packages to SQL Server 2005 Integration Services (SSIS) for a large electronic retailer. One source of information for the data warehouse was from TLog files, which contained data from point-of-sale cash registers in stores. TLog files store data in packed decimal format. To extract and transform this data into a format suitable for loading into the relational data warehouse tables, it was necessary to unpack and parse the files. To achieve this, custom pipeline components were implemented to read and transform the data by extending the SSIS object model. This paper discusses what we learned while implementing the pipeline components for extracting and parsing data from the TLog parser

2005-12-27

2,067 reads

Blogs

Rolling Back a Broken Release

By

We had an interesting discussion about deployments in databases and how you go forward...

A bespoke reporting solution doesn’t have to cost the earth

By

You could be tolerating limited reporting because there isn’t an off the shelf solution...

Presenting with Visual Studio Code

By

A while back I wrote a quick post on setting up key mappings in...

Read the latest Blogs

Forums

Cross-DB Ownership Chaining problem

By Johan Bijnens

We want to setup a gateway db to host stored procedures which use tables...

Lots of FKs

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Lots of FKs

Real-time On-prem SQL Server Data in Excel – Over the Internet

By Cláudio Tereso

Comments posted to this topic are about the item Real-time On-prem SQL Server Data...

Visit the forum

Question of the Day

Lots of FKs

In SQL Server 2025, what are the most outgoing and incoming FK references a table can have?

See possible answers