diponkar.paul

Diponkar Paul has 14 years of experience in the IT industry and specialized in the Business Intelligence domain, currently working as a Manager of Data Engineering at OMERS, Canada. Throughout his career, he designs and developed medium and large data warehouses and delivered business-critical analytical solutions. He was leading the Toronto PASS Group; which is part of the world's largest data professional community named PASS and Lead of Toronto Data Professionals Community. He got a diverse background and has been working and studying in different Geo locations including Canada, Denmark, the UK, and Sweden. He writes a technical blog and speaks, he loves to share his learning within the community.
  • Interests: Data, Travel, Swimming, Cycling, cross-country skate
  • Blog: http://allaboutdata.ca/
  • Jobs: Senior Data Engineer
  • Skills: Big data, Business Intelligence, Database, ETL, ELT, Data Warehouse

SQLServerCentral Article

Handling Row-level Errors in ADF Data Flows

If you are working with ADF (Azure Data Factory) data flows, then you may have noticed there was a new feature released in November 2020, which is useful to capture any error while inserting/updating the records in a SQL database. This article will describe how to setup the error row handling feature and why it's […]

(1)

You rated this post out of 5. Change rating

2021-06-16 (first published: )

5,667 reads

SQLServerCentral Article

Step by Step Guide to Install Jupyter Notebook

Whether you work as a Data Engineer or a Data Scientist, a Jupyter Notebook is a helpful tool. One of the projects I was working required a comparison of two parquet files. This is mainly a schema comparison, not a data comparison. Though the two .parquet were created from two different sources, the outcome should […]

You rated this post out of 5. Change rating

2021-05-17

5,504 reads

SQLServerCentral Article

Dynamically Add a Timestamp To Files in Azure Data Factory

This article will describe how to add your local timestamp at the end of the each file in Azure Data Factory (ADF). In general, ADF gets a UTC timestamp, so we need to convert the timestamp from UTC to EST, since our local time zone is EST. For example, if the input Source file name […]

(2)

You rated this post out of 5. Change rating

2021-04-22

30,249 reads

SQLServerCentral Article

How to simulate Case statement in Azure Data Factory (ADF) compared with SSIS?

This post will describe how you use a CASE statement in Azure Data Factory (ADF). If you are coming from SSIS background, you know a piece of SQL statement will do the task. However let's see how do it in SSIS and the very same thing can be achieved in ADF. Problem statement For my […]

(2)

You rated this post out of 5. Change rating

2020-11-12

22,486 reads

Blogs

Cost Visibility: Tracking and Analysing Your Cloud Spend

By

One of the biggest challenges I’ve faced in cloud operations is maintaining clear visibility...

Whiling away an afternoon, thinking

By

I come to Heathrow often. Today is likely somewhere close to 60 trips to...

Black Box vs. Gray Box vs. White Box Testing

By

If your organization is spending money, then meaningful results are a must. Pen testing...

Read the latest Blogs

Forums

Fun with JSON II

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Fun with JSON II

Changing Data Types

By Steve Jones - SSC Editor

Comments posted to this topic are about the item Changing Data Types

Answering Questions On Dropped Columns

By Cláudio Silva

Comments posted to this topic are about the item Answering Questions On Dropped Columns

Visit the forum

Question of the Day

Fun with JSON II

I have some data in a table:

CREATE TABLE #test_data
(
    id INT PRIMARY KEY,
    name VARCHAR(100),
    birth_date DATE
);

-- Step 2: Insert rows  
INSERT INTO #test_data
VALUES
(1, 'Olivia', '2025-01-05'),
(2, 'Emma', '2025-03-02'),
(3, 'Liam', '2025-11-15'),
(4, 'Noah', '2025-12-22');
If I run this query, how many rows are returned?
SELECT t1.[key] AS row,
       t2.*
FROM OPENJSON(
     (
         SELECT t.* FROM #test_data AS t FOR JSON PATH
     )
             ) t1
    CROSS APPLY OPENJSON(t1.value) t2;

See possible answers