diponkar.paul

Diponkar Paul has 14 years of experience in the IT industry and specialized in the Business Intelligence domain, currently working as a Manager of Data Engineering at OMERS, Canada. Throughout his career, he designs and developed medium and large data warehouses and delivered business-critical analytical solutions. He was leading the Toronto PASS Group; which is part of the world's largest data professional community named PASS and Lead of Toronto Data Professionals Community. He got a diverse background and has been working and studying in different Geo locations including Canada, Denmark, the UK, and Sweden. He writes a technical blog and speaks, he loves to share his learning within the community.
  • Interests: Data, Travel, Swimming, Cycling, cross-country skate
  • Blog: http://allaboutdata.ca/
  • Jobs: Senior Data Engineer
  • Skills: Big data, Business Intelligence, Database, ETL, ELT, Data Warehouse

SQLServerCentral Article

Handling Row-level Errors in ADF Data Flows

If you are working with ADF (Azure Data Factory) data flows, then you may have noticed there was a new feature released in November 2020, which is useful to capture any error while inserting/updating the records in a SQL database. This article will describe how to setup the error row handling feature and why it's […]

(1)

You rated this post out of 5. Change rating

2021-06-16 (first published: )

5,644 reads

SQLServerCentral Article

Step by Step Guide to Install Jupyter Notebook

Whether you work as a Data Engineer or a Data Scientist, a Jupyter Notebook is a helpful tool. One of the projects I was working required a comparison of two parquet files. This is mainly a schema comparison, not a data comparison. Though the two .parquet were created from two different sources, the outcome should […]

You rated this post out of 5. Change rating

2021-05-17

5,499 reads

SQLServerCentral Article

Dynamically Add a Timestamp To Files in Azure Data Factory

This article will describe how to add your local timestamp at the end of the each file in Azure Data Factory (ADF). In general, ADF gets a UTC timestamp, so we need to convert the timestamp from UTC to EST, since our local time zone is EST. For example, if the input Source file name […]

(2)

You rated this post out of 5. Change rating

2021-04-22

30,181 reads

SQLServerCentral Article

How to simulate Case statement in Azure Data Factory (ADF) compared with SSIS?

This post will describe how you use a CASE statement in Azure Data Factory (ADF). If you are coming from SSIS background, you know a piece of SQL statement will do the task. However let's see how do it in SSIS and the very same thing can be achieved in ADF. Problem statement For my […]

(2)

You rated this post out of 5. Change rating

2020-11-12

22,434 reads

Blogs

2025 Wrapped for Steve

By

I’ve often done some analysis of my year in different ways. Last year I...

The Book of Redgate: Spread across the world

By

This was Redgate in 2010, spread across the globe. First the EU/US Here’s Asia...

Merry Christmas

By

Today is Christmas and while I do not expect anybody to actual be reading...

Read the latest Blogs

Forums

SQL Server 2019 - Agent job PowerShell step issue

By Pete Bishop

I have a couple of SQL Agent job steps which run PowerShell commands of...

Database security permissions save script

By Srinivas Merugu

Comments posted to this topic are about the item Database security permissions save script

Database backup job steps

By Pete Bishop

I have a SQL Agent job for backing up a set of Analysis Services...

Visit the forum

Question of the Day

The Large Encoded Value

I want to use the new BASE64_ENCODE() function in SQL Server 2025, but return a string that isn't large type. What is the longest varbinary string I can pass in and still get a varchar(8000) returned?

See possible answers