External Article

Why Would I Ever Need to Partition My Big ‘Raw’ Data?

Whether you are running an RDBMS, or a Big Data system, it is important to consider your data-partitioning strategy. As the volume of data grows, so it becomes increasingly important to match the way you partition your data to the way it is queried, to allow 'pruning' optimisation. When you have huge imports of data to consider, it can get complicated. Bartosz explains how to get things right; not perfect but wisely.

External Article

How to Start Big Data with Apache Spark

It is worth getting familiar with Apache Spark because it a fast and general engine for large-scale data processing and you can use you existing SQL skills to get going with analysis of the type and volume of semi-structured data that would be awkward for a relational database. With an IDE such as Databricks you can very quickly get hands-on experience with an interesting technology.

Blogs

Houston AI-Lytics 2026–Powerpoint Slides

By

Thanks to everyone for attending my session on running a Local LLM. If you...

The Book of Redgate: Do the Right Things

By

I do believe that Redgate has been very customer focused since it’s inception. I’ve...

GenAI vs Dashboards: Not the Same (And Never Will Be)

By

There’s a question I’ve been hearing more and more lately, especially as Copilot, Fabric,...

Read the latest Blogs

Forums

Warning: Could not open global shared memory to communicate with performance DLL

By water490

Hi everyone I am getting below warning when I run SSIS: Warning: 0x80049304 at...

Let's Talk Community Events!

By Pat Wright

Comments posted to this topic are about the item Let's Talk Community Events!

that one limitation in replication

By stan

Hi as shown below a replication target requires a primary key.  if we want...

Visit the forum

Question of the Day

Detecting Deadlocks

By default, how often is the SQL Server Database Engine checking for deadlocks?

See possible answers