Skip to main content


Immutable blobs inside Azure Storage (WORM)

Recent posts

How to calculate the cost of CI/CD on top of Azure VSTS

In this post, we tackle the cost estimation of a CI/CD system that is hosted on top of Visual Studio Team Services (VSTS).

We are starting from an on-premises system that has a CI/CD system build on top of Jenkins, which deploys the web application on a local IIS. Beside web application, we have for each project an SQL Server Database that can reach up to 100GB. The source control repository is GitHub, and for task management and tracking, we have JIRA.
In the current environment, we have around 20 projects that need such a migration, but the cost estimation is done project base, allowing us to have a better forecast for each project. We don’t take into consideration any other dependencies that a specific project could have like Windows Services.
Most of the client is already using Azure to host their production environments for their web application. Going into the cloud is more than acceptable for them. Beside this remember that you have the code inside GitHub.
In some cases,…

Using an external email provider with Episerver DXC environment inside Azure

In this post, we talk about different options that we have we need to integrate an email service with an eCommerce application developed using Episerver and hosted inside DXC. As a side note, DXC it is SaaS on top in Microsoft Azure, where clients can host their web applications developed on top of Episerver. More about the DXC environment will be covered in another post.
What we what to achieve? We want to be able to integrate an external email service provider to be able to send emails to our clients for marketing purposes. Our web application it is already hosted inside DXC, and even if we could write custom code that can run inside DXC to communicate with the email service provider, there are some limitations that we need to be aware. The authentification and authorization mechanism offered by the email service provider it is based on IP whitelisting. Only the IPs from that whitelist are allowed to make calls to the service and send emails.
Limitations At this moment in time, Microso…

Data moving to/from Azure Storage using Azure Data Factory and Copy Activity

This post covers how we can use Azure Data Factory to copy our data from one location to another. Even if the copy activity might sound like a basic task that is not complex or challenging when you need to move 10 TB or 50 TB of data from one storage to another, things are becoming interesting. Reliability, resume and continue, task parallelization, data consistency are becoming essential for your project. There are multiple solutions on the market that helps you to move data to and from Azure or between different Azure locations. Today we will focus on Azure Data Factory and how it can help us to our job better. Azure Data Factory it is a service created to help us to work with data when we need to store, process and move it, without having to things about anything else.
Inside Azure Data Factory there we have an activity called Copy Activity that can be used to move (copy) data from source to another. When we talk about Azure Data Factory, we call: Source the repository from where we r…

Azure Design patterns: Self-Healing & Transient fault handling

I will start a series of posts about the core design patterns that you need to take into consideration when you start to work with Microsoft Azure and the cloud in general.
These principles are important in a cloud application, even if most of them are known from classical on-premises development. Even if most of them are known, we don’t apply them for on-premises system all the time.
In a cloud environment, it is possible for example to have for 10ms a connectivity problem with another service that you are using. It means that you would need a retry policy in-place that would automatically try to reconnect to the service.
Cases like this need to be covered by your application that is running in the cloud. Take into consideration that you need to take all of this into consideration for lift and shift cases, too.

Self-Healing & Transient fault handling
The today post it is dedicated to self-healing topic. This is not SF, and you don’t need to be Netflix to have a self-healing mechanism.…

Why to use Azure Databricks to run your Apache Spark Notebooks

A few months ago (almost one year) a new flavor of Apache Spark appeared on Microsoft Azure. Azure Databricks it is just a platform optimized for Azure, where Apache Spark can run.
The consumer does not need anymore to configure the Apache Spark cluster (VM creation, configuration, network, security, storage and many more). Azure Databricks already has a cluster that is configured and ready to be used. We can focus on our application and business requirements and less on the infrastructure part.

Capabilitiesand Features
All the features that we have inside Apache Spark can also be found inside Azure Databricks. The same version of Spark that you have on-premises runs on top of Azure Databricks, the only difference is at the infrastructure level, where you already have the system preconfigured. As a user, you are free to scale up or down your cluster in a 'drag and drop' manner, without having to pay attention or do anything else.
All the 5 main components of Spark can be found …

Azure Dev Spaces - A free tool for developing app on top of AKS

In this post, we will talk about Azure Dev Spaces and how it can improve the developing and support experience when you are working with AKS (Azure Kubernetes Service).

Working with containers and AKS it is a great experience when you develop the application on your local machine, or you are in production and everything works as expected.
The friction appears in the moment when you have some issues with your application, and you want to debug it. Recreating a dedicated env. just for you can be complicated, time-consuming and expensive.  This is where Azure Dev Spaces comes to fill the gap. It is enabling us to deploy and to remote debugging of our services directly inside the AKS cluster.

Directly from our preferred IDE we can deploy our container inside AKS and do a remote debugging session to it, adding breakpoint directly in Visual Studio.

How is this possible?
Azure Dev Spaces tools that are installed on your local machine are creating a secure tunnel between your device and AKS. T…