2025-12-22
560 reads
2025-12-22
560 reads
2024-07-12
606 reads
A column level check constraint may be used to prevent insertion of data containing characters within a specific ASCII range.
2016-12-12 (first published: 2016-11-30)
2,323 reads
Script finds char types columns defined to allow NULL values but have non NULL values and generates the ALTER statements for the changes.
2016-02-22 (first published: 2016-02-08)
1,018 reads
In SQL, you can express the logic of what you want to accomplish without spelling out the details of how the database should do it. Nowhere is this more powerful than in constraints. In this introduction to Declarative SQL, Joe Celko demonstrates how you can write portable code that performs well and executes some complex logic, merely by creating unique constraints.
2015-12-15
6,149 reads
2015-11-18 (first published: 2015-10-29)
1,006 reads
In which Phil Factor illustrates in TSQL how it is possible to use foreign key constraints to enforce data rules, and illustrates some surprising consequences of using cascading.
2015-03-30
3,057 reads
When you're importing data into an RDBMS in bulk and an exception condition is raised because of a constraint violation, you generally need to fix the problem with the data and try again. The error won't tell you which rows are causing the violation. What if you've thousands of rows to search when it happens? There are solutions, writes William Sisson.
2014-10-16
8,821 reads
The database developer can, of course, throw all errors back to the application developer to deal with, but this is neither kind nor necessary. How errors are dealt with is dependent on the application, but the process itself isn't entirely obvious.
2012-07-10
3,666 reads
We are storing large text and URLs that are over 900 bytes in some of our tables and have a requirement to enforce uniqueness in those columns. But SQL Server has a limitation that index size can't be over 900 bytes. How do I enforce uniqueness in these columns and is it possible to achieve this in SQL Server 2005 and above? What are my different options to solve this problem? I heard that we can use CHECKSUM to create a hash, but is it possible to avoid collisions in the hash value as we are storing millions of rows?
2009-11-05
3,006 reads
By Brian Kelley
If your organization is spending money, then meaningful results are a must. Pen testing...
By HeyMo0sh
In my experience, FinOps success has never been just about tools or dashboards. It...
By HeyMo0sh
As a DevOps person, I know that to make FinOps successful, you need more...
Comments posted to this topic are about the item Restoring On Top II
Comments posted to this topic are about the item SQL Art 2: St Patrick’s...
Comments posted to this topic are about the item Breaking Down Your Work
I have a database, DNRTest, that has a number of tables and other objects in it. The other day, I was trying to mock up a test and ran this code on the same server:
-- run yesterday CREATE DATABASE DNRTest2 GO USE DNRTest2 GO CREATE TABLE NewTable (id INT) GOToday, I realize that I need a copy of DNRTest for another mockup, and I run this:
-- run today USE Master BACKUP DATABASE DNRTest TO DISK = 'dnrtest.bak' GO RESTORE DATABASE DNRTest2 FROM DISK = 'dnrtest.bak' WITH REPLACEWhat happens? See possible answers