Additional Articles


External Article

Best Practices – When Using FILESTREAM Feature of SQL Server 2008

In SQL Server 2008 one can store BLOBs (e.g. Images, video, Word, Excel, PDF, MP3, etc) in the NTFS file system rather than in a database file. This can be done by using the new FILESTREAM feature which was introduced in SQL Server 2008. In this tip we will take a look at some of the best practices which a database administrator can follow to get the best performance when using the FILESTREAM feature of SQL Server 2008.

2009-11-11

4,240 reads

Technical Article

What a Data Warehouse is Not

Recently, I was at a conference, and I heard the following discussion about what a data warehouse was. One person suggested that a data warehouse was really all the old legacy systems connected by software that could access the data. By calling such a contraption a data warehouse, the organization could avoid having to do the hard and complex work of integration. There are so many problems with this federated approach to a data warehouse that they are almost not worth repeating here. But (once again!) here goes.

2009-11-09

6,637 reads

Technical Article

Using T-SQL CROSS APPLY and OUTER APPLY

That said, even though my personal focus is pretty much OO, I still miss stuff. SQL Server 2005 came out with the CROSS APPLY and OUTER APPLY operators and I have just started learning how to use APPLY in the last month or so. When I think I have it figured out, I am fortunate enough that some of you are interested in reading about my understanding of the technology.

2009-11-06

7,434 reads

External Article

SQL Server Unique Constraints for Large Text Columns

We are storing large text and URLs that are over 900 bytes in some of our tables and have a requirement to enforce uniqueness in those columns. But SQL Server has a limitation that index size can't be over 900 bytes. How do I enforce uniqueness in these columns and is it possible to achieve this in SQL Server 2005 and above? What are my different options to solve this problem? I heard that we can use CHECKSUM to create a hash, but is it possible to avoid collisions in the hash value as we are storing millions of rows?

2009-11-05

3,006 reads

Blogs

Resetting on the AI hype train

By

There's a great article from MIT Technology Review about resetting on the hype of...

A New Word: Etherness

By

etherness – n. the wistful feeling of looking around a gathering of loved ones,...

Vibe Coding a Login Tracking System

By

A customer was asking about tracking logins and logouts in Redgate Monitor. We don’t...

Read the latest Blogs

Forums

The Max PK Length

By Steve Jones - SSC Editor

Comments posted to this topic are about the item The Max PK Length

My experience using the GitHub Copilot in SSMS 22

By Daniel Calbimonte

Comments posted to this topic are about the item My experience using the GitHub...

The Microsoft SQL Year in Review

By Steve Jones - SSC Editor

Comments posted to this topic are about the item The Microsoft SQL Year in...

Visit the forum

Question of the Day

The Max PK Length

If I create a multiple column Primary Key constraint, what is the most number of bytes I can include in the constraint?

See possible answers