Big Data

External Article

Why Would I Ever Need to Partition My Big ‘Raw’ Data?

  • Article

Whether you are running an RDBMS, or a Big Data system, it is important to consider your data-partitioning strategy. As the volume of data grows, so it becomes increasingly important to match the way you partition your data to the way it is queried, to allow 'pruning' optimisation. When you have huge imports of data to consider, it can get complicated. Bartosz explains how to get things right; not perfect but wisely.

2016-11-22

3,345 reads

External Article

How to Start Big Data with Apache Spark

  • Article

It is worth getting familiar with Apache Spark because it a fast and general engine for large-scale data processing and you can use you existing SQL skills to get going with analysis of the type and volume of semi-structured data that would be awkward for a relational database. With an IDE such as Databricks you can very quickly get hands-on experience with an interesting technology.

2016-11-18

3,131 reads

External Article

The End of Big Data

  • Article

What is next for big data? Some experts claim that data "volumes, velocity, variety and veracity" will only increase over time, requiring more data storage, faster machines and more sophisticated analysis tools. However, this is short-sighted, and does not take into account how data degrades over time. Analysis of historical data will always be with us, but generation of the most useful analyses will be done with data we already have. To adapt, most organizations must grow and mature their analytical environments. Lockwood Lyon shares the steps they must take to prepare for the transition.

2016-06-03

10,764 reads

External Article

Big Data Architecture

  • Article

The next few years will be critical for the information technology staff, as they attempt to integrate and manage multiple, diverse hardware and software platforms. In this article, Lockwood Lyon addresses how to meet this need, as users demand greater ability to analyze ever-growing mountains of data, and IT attempts to keep costs down.

2016-05-09

5,553 reads

Blogs

ISACA AI Material/Exam Prep Discount (May 18 – June 30, 2026)

By

If you are considering any of the ISACA AI certs like the Advanced Artificial...

A Fabric solution can be very cost effective

By

Are you currently using Microsoft Fabric or considering migrating to it? If so, there...

Track SQL Server Configuration Changes Using the Error Log

By

Track SQL Server Configuration Changes Using the Error Log If you...

Read the latest Blogs

Forums

Designing SQL Server ETL Pipelines That Don't Break at Scale

By SQL Expert

Comments posted to this topic are about the item Designing SQL Server ETL Pipelines...

Detecting Deadlocks Quickly

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Detecting Deadlocks Quickly

SQL Server Showing Out of Space But Plenty of Drive Space

By AMCity

We have a SQL Server installed. We have a 500GB drive for the database....

Visit the forum

Question of the Day

Detecting Deadlocks Quickly

In the Database Engine, when a deadlock is detected, what does the detection interval shrink to (in time)?

See possible answers