Azure

Download a file from DBFS – Databricks to the local machine

In this post, we will learn how we can download a file from DBFS i.e. Databricks File System to the Local machine. DBFS is the File system that Databricks uses to store its files. It is a distributed file system mounted into a Databricks workspace and it is available on Databricks clusters. To demonstrate how […]

Download a file from DBFS – Databricks to the local machine Read More »

Create an Azure Function in Java using IntelliJ IDEA

Azure Functions are serverless compute that help orchestrate and automate complex problems in an event-driven way. We can choose any supported language to create and deploy an Azure Function in the cloud. The supported language as of date is C#, Java, Python, JavaScript, F#, Powershell, and TypeScript. This post will learn how we can create

Create an Azure Function in Java using IntelliJ IDEA Read More »

Use HDFS API to read Azure Blob files in Databricks

Databricks provides a wrapper file system API named DBFS (Databricks File System) to perform any file-level operation such as read, write, move, delete, rename, etc. However, sometimes we may need to read the underlying file system objects directly without using the DBFS wrapper APIs. To do so, we can use HDFS APIs available through py4j

Use HDFS API to read Azure Blob files in Databricks Read More »

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database testing using tSQLt – Part 4

In the previous posts, we have created a Continuous integration and a Conntinuous Deployment pipeline for a SQL Server database using the Azure DevOps server. Also, we have demonstrated how we can set up the cross-database dependency for a SQL Server database project in the Azure DevOps pipeline. Below are the links in case you

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database testing using tSQLt – Part 4 Read More »

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database project dependency – Part 3

Previously, we have created an Azure DevOps Continuous Integration (CI) and Continuous deployment (CD) pipelines to independently deploy a SQL Server database project. However, in an enterprise data warehouse environment, the databases are mostly dependent on other databases. Because, in DWH, we access objects from multiple databases (like accessing the staging layer objects into the

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database project dependency – Part 3 Read More »

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database CD – Part 2

In the previous post, we have created a CI (Continous Integration) pipeline for a SQL Server database project. For this demo, as like our previous demo, we will be using a SQL Server instance running on an on-prem machine along with a locally installed Azure DevOps Server. Please note that we are not using the

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database CD – Part 2 Read More »

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database CI – Part 1

In this post, we are going to discuss how we can enable continuous integration and continuous deployment for a SQL Server Database project using Azure DevOps Server. For this demo, we will be using a SQL Server instance running on-prem along with a locally installed Azure DevOps Server. Continuous integration and continuous delivery (in short

Continuous Integration and Continuous Deployment (CI/CD) – SQL Server Database CI – Part 1 Read More »

Handling special characters in Hive (using encoding properties)

In case we are reading a text file in a Hive table which contains non-English characters and we are not using the appropriate text encoding, these non-English characters might be loaded as junk symbols (like boxes – �). To get these characters in their original form, we need to use the correct character encoding. In this

Handling special characters in Hive (using encoding properties) Read More »

Skip header and footer rows in Hive

In this post “Skip header and footer rows in Hive“, we are going to learn that how we can ignore few header and footer records in Hive without loading or reading these records in another table or in a view temporarily. If you want to read more about Hive, visit my post “Preserve Hive metastore in

Skip header and footer rows in Hive Read More »