Additional Articles


External Article

Usage-Based Optimization in Analysis Services 2005

In this lesson, we revisit usage-based optimization, a subject that we undertook in my article MSAS Administration and Optimization: Simple Cube Usage Analysis, in September of 2003, and MSAS Administration and Optimization: Toward More Sophisticated Analysis in October of 2003. In the earlier articles, we discovered that, among several tools that Microsoft SQL Server 2000 Analysis Services offered us to assist in the maintenance and optimization of our cubes, two of these tools, the Usage Analysis Wizard and the Usage-Based Optimization Wizard, leveraged the usage-based optimization features of Analysis Services.

2006-01-10

1,810 reads

Technical Article

PART 8: ESTIMATION

Data modeling is no doubt one of the most important and challenging aspects of developing, maintaining, augmenting and integrating typical enterprise systems. More than 90% of functionality of enterprise systems is centered round creating, manipulating and querying data. It therefore stands to reason that individuals managing enterprise projects should leverage on data modeling to execute their projects successfully and deliver not only capable and cost effective but also maintainable and extendable systems

2006-01-09

2,421 reads

Technical Article

THE ROLE OF DATA DICTIONARY

IT leaders from enterprises around the world have realized that acquiring and implementing commercial off-the-shelf (COTS) application packages is a smarter thing to do, than the traditional approach of developing home-grown applications. This may not be the fit-for-all solution, but it makes sense from a cost, time to market and resource standpoint. Selection and acquisition of the software package is just the first step. How the software package brings value through successful implementation, is critical to the business. Internal groups as well as industry analysts will closely watch COTS package implementation in the organization.

2006-01-05

2,981 reads

Technical Article

Using ITWIZ.EXE to Automate Index Analysis

This process uses ITWIZ.EXE the command line version of the Index Tuning Wizard. SQL trace file information is stored in a table on production then transferred to a non-production server. ITWIZ is then run against the transferred trace data resulting in a UNICODE SQL script. This script has index creation and deletion information as well as script parameters in a remarks section.

2005-12-29

1,973 reads

Blogs

What is ALM in Fabric?

By

As someone who’s worked with data for over 20 years and with many cloud...

The Most Successful Startups in 2025 — And What They Have in Common

By

2025 belongs to the AI startups. If you peek into the tech headlines, you’ll...

Blog a Day – Day 1: History of AI

By

it has been a year since i have not written much on the blog...

Read the latest Blogs

Forums

A Quick Restore

By Steve Jones - SSC Editor

Comments posted to this topic are about the item A Quick Restore

Guarding Against SQL Injection at the Database Layer (SQL Server)

By tedo

Comments posted to this topic are about the item Guarding Against SQL Injection at...

Ola Hallengren Index Optimize Maintenance can we have data compression = page

By JSB_89

I have a quick question on Ola Hallengren Index Optimize Maintenance . Do we...

Visit the forum

Question of the Day

A Quick Restore

While doing some testing of an application, I wanted to reset my environment after doing some testing with this code:

USE DNRTest

BACKUP DATABASE DNRTest TO DISK = 'dnrtest.bak'
GO
/*
Bunch of stuff tested here
*/RESTORE DATABASE DNRTest FROM DISK = 'dnrtest.bak' WITH REPLACE
What happens if this runs, assuming the "bunch of stuff" isn't anything affecting the instance.

See possible answers