External Article

Scala and Apache Spark in Tandem as a Next-Generation ETL Framework

Scala and Apache Spark might seem an unlikely medium for implementing an ETL process, but there are reasons for considering it as an alternative. After all, many Big Data solutions are ideally suited to the preparation of data for input into a relational database, and Scala is a well thought-out and expressive language. Krzysztof Stanaszek describes some of the advantages and disadvantages of a scala-based approach to implementing and testing an ETL solution.

Blogs

Advice I Like: Rewards from Work

By

The greatest rewards come from working on something that nobody has words for. If...

Overcoming Challenges: Navigating Common Pitfalls in FinOps Adoption

By

Working in DevOps, I’ve seen FinOps do amazing things for cloud cost control, but...

Why your data still can’t answer a simple question 

By

Every organization I talk to has the same problem dressed up in different clothes....

Read the latest Blogs

Forums

The day-to-day pressures of a DBA team, and how we can work smarter with automation and AI

By Terry Jago

Comments posted to this topic are about the item The day-to-day pressures of a...

Analysis Services Model w/ Direct Query and (Default Veritpaq)

By Archivist

Analysis Services (either the integrated workspace in Power BI or on a SQL Server)...

Don't Panic

By Grant Fritchey

Comments posted to this topic are about the item Don't Panic

Visit the forum

Question of the Day

Identities and Sequences IV

When thinking about the identity property and sequence objects, which of these can be used with numeric and decimal data types?

See possible answers