Dimensional Modeling Case Study Part 3 - Dynamic Days Dimension
In this next installment of the date dimension series, learn how to create a table that supports different types of banding.
2025-06-13
2,064 reads
In this next installment of the date dimension series, learn how to create a table that supports different types of banding.
2025-06-13
2,064 reads
Real-time data ingestion has become essential for modern analytics and operational intelligence. Organizations across industries need to process data streams from IoT sensors, financial transactions, and application events with minimal latency. Snowflake offers two robust approaches to meet these real-time data needs: Snowpipe for near-real-time file-based streaming and Direct Streaming via Snowpark API for true real-time data integration.
2025-06-13
Validate all of your settings and be prepared to make some changes during your migration process. Most of the incompatible options make sense when you think about the purpose of SQL MI – it is controlled by Microsoft. Hardware settings, local file access, high-availability settings, and auditing are configured differently or completely disabled.
2025-06-11
Learn a few ways to improve performance in your Azure SQL Databases through better indexing, partitioning, and columnstore index consideration.
2025-06-09
3,428 reads
The new JSON field type and new functions in Azure SQL brings a new world of possibilities for JSON manipulation in SQL Databases. These features are slated to be a part of SQL Server 2025 as well. Let’s analyze some practical uses of these features
2025-06-09
SQLAutomate is a lightweight PowerShell tool that simplifies SQL Server deployment by automating both installation and post-installation configuration. Whether you're setting up one instance or dozens, this guide shows how to use SQLAutomate to save time and reduce manual work.
2025-06-06
10,186 reads
There are several things that you can do to improve performance by throwing more hardware at the problem, but usually the place you get the most benefit from is when you tune your queries. One common problem that exists is the lack of indexes or incorrect indexes and therefore SQL Server has to process more data to find the records that meet the queries criteria. These issues are known as Index Scans and Table Scans.
2025-06-06
In this next article, we are going to explore how to install, configure and use the command line to manage a couple different Fabric Lakehouse’s.
2025-06-04
1,567 reads
Purging data from a table is a common database maintenance task to prevent it from growing too large or to stay in compliance with data retention. When dealing with small amounts of data, this can be accomplished by a simple delete with no issues; however, with larger tables, this task can be problematic. Deleting records requires a lock that can block other processes from writing or even reading the data (depending on your isolation level). In this article I will share a technique I have used to work with some very large tables.
2025-06-04
What DMVs were added to Azure SQL Databases, and which DMVs are missing from Azure SQL Databases but still exist on SQL Server 2019 or other versions?
2025-06-02
By Vinay Thakur
Continuing from Day 4 where we learned Encoder, Decoder, and Attention Mechanism, today we...
By Vinay Thakur
Continuing from Day 3 where we covered LLM models open/closed and their parameters, Today...
By Steve Jones
One of the nice things about Flyway Desktop is that it helps you manage...
I'm fairly certain I know the answer to this from digging into it yesterday,...
Hi Team, I am trying to refresh the Azure Synapse Dedicated pool from production...
hi everyone I am not sure how to write the query that will produce...
I have some data in a table:
CREATE TABLE #test_data
(
id INT PRIMARY KEY,
name VARCHAR(100),
birth_date DATE
);
-- Step 2: Insert rows
INSERT INTO #test_data
VALUES
(1, 'Olivia', '2025-01-05'),
(2, 'Emma', '2025-03-02'),
(3, 'Liam', '2025-11-15'),
(4, 'Noah', '2025-12-22');
If I run this query, how many rows are returned?
SELECT *
FROM OPENJSON(
(
SELECT t.* FROM #test_data AS t FOR JSON PATH
)
) t; See possible answers