Skip to main content

(Part 1) Azure Storage - Cold Tier | Things that we should know

In this post we will talk about things that we need to consider if we want to use Cold storage of Azure Storage. This post we’ll be followed by another one that will present use cases when cold storage tier is useful and how we should integrated it in our system.
(Part 2) Azure Storage - Cold Tier | Business use cases where we could use it

Overview
Cold storage tier of Azure Storage is a new tier that is available for Azure Storage for data that needs to be persisted for long time with minimum costs. This tier should be used when data is not accessed frequently. In comparison with Hot tier, cold storage latency can be a little bit higher, but still it is in milliseconds.
Azure Storage contains multiple types of storages like Azure Queues, Azure Tables, Azure Blobs (block blob, append blob and page blob). Unfortunately, their tier is available only for block blob and append blob. It means that we cannot store an Azure Table or a Page Blob in the cold tier.
The Storage Account that allows us to have access to this storage tier also is only available when we create a new storage account and we specify at the account type “Blob Storage” and not “General Purpose”. Only in this way we will have access to the cold storage tier.

Things that we should take into account
There are some things that you should consider also. Once you create an Azure Storage account that has support for Cold/Hot tier you can switch between. Switching between this two tier is not free of cost. The cost of migration from one tier to another is equivalent with cost of reading all your data from that storage account.
Existing storage accounts that are General Purpose or old storage account (so called classical ones) cannot be change to Blob Storage kind. It means that for this storages we cannot have cold and hot tier functionality.

For this scenarios it is necessary to migrate all the content to specific storage accounts. This can be made easily using AzCopy tool, that will make the migration for you. This tools can be used directly from command prompt or write a small app on top of Azure Data Movement library that is using AzCopy behind the scene.
The API that can be used for block and append blob it is the same as for normal Storage Accounts. If you do such a migration that are no change inside your app code. Only the connection string that points to the Azure Storage account needs to be changed.

HOT vs COLD Tier
If we compare this two tiers, at SLA level there is a small different for the Availability SLA. Where from 99.99% (RA-GRS) availability for Hot tier, we have only 99.9 availability for Cold storage.
From cost perspective, HOT storage cost of storage is higher than the Cold one (more than double), but in the same time the transaction cost is almost double for Cold storage. This is an important thing that we need to take into account when we make the cost estimation. Also on top of this,  there are costs for reading and writing to cold tire in comparison with hot tier when you pay only for outbound data, that go outside the Azure Region
The latency is the same for both of them – expressed in milliseconds.


When you are using the Geo-Replication feature for Cold tier, the replication cost in the secondary region will be as you read the data from the main location.

In the next post we will take a look on use cases when we should use Cold tier and how the migration should be realizes.
(Part 2) Azure Storage - Cold Tier | Business use cases where we could use it

Comments

Popular posts from this blog

How to check in AngularJS if a service was register or not

There are cases when you need to check in a service or a controller was register in AngularJS.
For example a valid use case is when you have the same implementation running on multiple application. In this case, you may want to intercept the HTTP provider and add a custom step there. This step don’t needs to run on all the application, only in the one where the service exist and register.
A solution for this case would be to have a flag in the configuration that specify this. In the core you would have an IF that would check the value of this flag.
Another solution is to check if a specific service was register in AngularJS or not. If the service was register that you would execute your own logic.
To check if a service was register or not in AngularJS container you need to call the ‘has’ method of ‘inhector’. It will return TRUE if the service was register.
if ($injector.has('httpInterceptorService')) { $httpProvider.interceptors.push('httpInterceptorService&#…

ADO.NET provider with invariant name 'System.Data.SqlClient' could not be loaded

Today blog post will be started with the following error when running DB tests on the CI machine:
threw exception: System.InvalidOperationException: The Entity Framework provider type 'System.Data.Entity.SqlServer.SqlProviderServices, EntityFramework.SqlServer' registered in the application config file for the ADO.NET provider with invariant name 'System.Data.SqlClient' could not be loaded. Make sure that the assembly-qualified name is used and that the assembly is available to the running application. See http://go.microsoft.com/fwlink/?LinkId=260882 for more information. at System.Data.Entity.Infrastructure.DependencyResolution.ProviderServicesFactory.GetInstance(String providerTypeName, String providerInvariantName) This error happened only on the Continuous Integration machine. On the devs machines, everything has fine. The classic problem – on my machine it’s working. The CI has the following configuration:

TeamCity.NET 4.51EF 6.0.2VS2013
It seems that there …

[Post-Event] Codecamp Conference Cluj-Napoca - Nov 19, 2016

Last day I was invited to another Codecamp Conference, that took place in Cluj-Napoca. Like other Codecamp Conferences, the event was very big, with more than 1.000 participants and 70 sessions. There were 10 tracks in parallel, so it was pretty hard to decide at  what session you want to join.
It was great to join this conference and I hope that you discovered something new during the conference.
At this event I talked about Azure IoT Hub and how we can use it to connect devices from the field. I had a lot of demos using Raspberry PI 3 and Simplelink SensorTag. Most of the samples were written in C++ and Node.JS and people were impressed that even if we are using Microsoft technologies, we are not limited to C# and .NET. World and Microsoft are changing so fast. Just looking and Azure IoT Hub and new features that were launched and I'm pressed (Jobs, Methods, Device Twin).
On backend my demos covered Stream Analytics, Event Hub, Azure Object Storage and DocumentDB.

Title:
What abo…