SQLServerCentral Article

Using Logging with the Delete Activity in Azure Data Factory

,

Introduction

In this article, we will discuss the log configuration in the Delete Activity in Azure Data Factory.

The Source File

I put a sample file, Sales.csv, in the folder, adfdemo/Files.

We will use the logging settings in the Delete activity:

Let's configure the source dataset, I added the linked service for Azure blob storage., then edited the source file path, which is adfdemo/Files/Sales.csv.

Now I will configure the log settings as below:

  1. Enable logging: must be enabled
  2. Logging account linked service: Linked service for the logging
  3. Folder path: Folder path to creating the log file.

Let's run the pipeline:

 

The source file is deleted from the source folder:

 

A new folder is created in the log folder:

 

A new log file is created in the log folder:

 

Now I open the log file, it has four columns as below:

  • Name - Name of the deleted file or folder.
  • Category - Specifies whether it is a file or a folder.
  • Status - Delete status.
  • Error - Error message if any.

Every time the delete activity deletes a file or folder, it creates a new logfile within a folder.

Now we will run the pipeline once again with the empty file in the source folder:

Let's run the pipeline:

The log file is created:

I opened the log file and it is showing the error message in the error column - File does not exist because there was no source file in the folder.

 

Conclusion

In this article, we discussed how to use a log file in the delete activity. The log file is very useful to check the delete status for the specified files or folders.

 

Rate

Share

Share

Rate