Big Data

SQL Server Integrates Hadoop and Spark out-of-the box: The Why?

  • Article

Introduction Microsoft announced in September 2018 that SQL Server 2019, which is now in preview, will have a Big Data Cluster deployment option. This is a Big-Data-capable SQL Server with elastic scale and extended Artificial Intelligence (AI) capabilities, mostly as a result of deep integration of Hadoop and Spark out-of-the-box.  The new SQL Server Big […]

4.67 (6)

2019-09-09

4,512 reads

Why Would I Ever Need to Partition My Big ‘Raw’ Data?

  • Article

Whether you are running an RDBMS, or a Big Data system, it is important to consider your data-partitioning strategy. As the volume of data grows, so it becomes increasingly important to match the way you partition your data to the way it is queried, to allow 'pruning' optimisation. When you have huge imports of data to consider, it can get complicated. Bartosz explains how to get things right; not perfect but wisely.

2016-11-22

3,345 reads

How to Start Big Data with Apache Spark

  • Article

It is worth getting familiar with Apache Spark because it a fast and general engine for large-scale data processing and you can use you existing SQL skills to get going with analysis of the type and volume of semi-structured data that would be awkward for a relational database. With an IDE such as Databricks you can very quickly get hands-on experience with an interesting technology.

2016-11-18

3,131 reads

The End of Big Data

  • Article

What is next for big data? Some experts claim that data "volumes, velocity, variety and veracity" will only increase over time, requiring more data storage, faster machines and more sophisticated analysis tools. However, this is short-sighted, and does not take into account how data degrades over time. Analysis of historical data will always be with us, but generation of the most useful analyses will be done with data we already have. To adapt, most organizations must grow and mature their analytical environments. Lockwood Lyon shares the steps they must take to prepare for the transition.

2016-06-03

10,764 reads

Big Data Architecture

  • Article

The next few years will be critical for the information technology staff, as they attempt to integrate and manage multiple, diverse hardware and software platforms. In this article, Lockwood Lyon addresses how to meet this need, as users demand greater ability to analyze ever-growing mountains of data, and IT attempts to keep costs down.

2016-05-09

5,553 reads

Blogs

T-SQL Tuesday Live–7 Apr 2020

By

Last week went a little crazy, and I learned something: don’t post a Zoom...

Getting Your dbatools Version–#SQLNewBlogger

By

Another post for me that is simple and hopefully serves as an example for...

SQL Server and VMware Resource Pools

By

Resource pools are used to hierarchically partition available CPU and memory resources, and are...

Read the latest Blogs

Forums

Tools for DISA's SQL Server STIGs

By aborelli1

Is it possible to request a review of a tool I created to automate...

bcp format file and import fixed width text file

By smattiko83

I'm trying to set up a format file using bcp but I'm unsure of...

insert random value from predefined list of possible values

By marchello

Hi all, My need is to insert random value from predefined list of possible...

Visit the forum

Ask SSC

SQL Server Q&A from the SQLServerCentral community

Get answers

Question of the Day

Linux Support

See possible answers