Build a Streaming Data Architecture with Apache Kafka and Zookeeper
In this article, we look at how to use Apache Kafka and Zookeeper to process and load streaming data.
2024-12-16
In this article, we look at how to use Apache Kafka and Zookeeper to process and load streaming data.
2024-12-16
This article examines how one can structure a pipeline for processing real-time data using Kafka and Informatica.
2023-04-26
4,646 reads
See a simple demonstration of how Kafka can stream events for changes to data in SQL Server tables.
2021-07-02 (first published: 2021-05-27)
16,808 reads
One feature that I have been waiting for years! The new announcement around optimize...
Following on from my last post about Getting Started With KubeVirt & SQL Server,...
By DesertDBA
I haven’t posted in a while (well, not here at least since I’ve been...
Comments posted to this topic are about the item Refactoring SQL Code, which is...
Comments posted to this topic are about the item The Read Committed Snapshot Isolation...
Comments posted to this topic are about the item Working with JSON/JSONB Data in...
I am currently working with Sql Server 2022 and AdventureWorks database. First of all, let's set the "Read Committed Snapshot" to ON:
use master; go alter database AdventureWorks set read_committed_snapshot on with no_wait; goThen, from Session 1, I execute the following code:
--Session 1 use AdventureWorks; go create table ##t1 (id int, f1 varchar(10)); go insert into ##t1 values (1, 'A');From another session, called Session 2, I open a transaction and execute the following update:
--Session 2 use AdventureWorks; go begin tran; update ##t1 set f1 = 'B' where id = 1;Now, going back to Session 1, what happens if I execute this statement?
--Session 1 select f1 from ##t1 where id = 1;See possible answers