Skip to main content

[IoT Home Project] Part 8 - Connecting to Azure Function and to a virtual heat pump

In this post we will discover how to:
  • Push content from Azure Stream Analytics to Azure Service Bus
  • Write an Azure Function in Node.JS that fetch data from Azure Service Bus Topic and push it to Azure Table
  • Develop a ASP.NET Core application that plays the role of a heating system that start/stop the heating in a house
  • Use temperature data collected from sensors connected to Raspberry PI to start/shop a heating system from a house
Previous post: [IoT Home Project] Part 7 - Read/Write data to device twin
GitHub source code: 

Push content from Azure Stream Analytics to Azure Service Bus
This step is the most simple one. We just need to Add a new output to Stream Analytics and specify the Service Bus Topic where we want to push data. After this, we'll need to update the Stream Analytic query by adding:
, where 'outputSensorDataTopic' is the name of the output that we created a few seconds ago.

Remarks: Don't forget to create a subscription for the topic. I send all the information to topic (not only temperature), because I plan in the future to add filters to each subscription and root data based on this parameters.

Azure Function to push data from Service Bus Topic to Azure Table
Offtopic: This step is done artificially in our case, we could forward content directly to Azure Table from Stream Analytics. I wanted to do this because the main scope is to play with as many Azure Services as possible. Additional to this, inside Azure Function we could aggregate information from other sources to decide if we want to start/stop the heating system.

Now, we will create a Azure Table in our storage with two rows, that contain current temperature and desired temperature. We could also add another row that contain a flag that specify to the system to start/stop the heating system, but for now we will not add it.
The table should look like this:
From Azure Portal, in the moment when we want need to create the Function itself we shall specify that for Trigger we want to have Service Bus Topic. For output we'll specify Azure Tables. At this step you need to specify connection information for Service Bus and Azure Table. Once you done this, you'll be ready to write your first Azure Function in Node.JS
  module.exports = function (context, myQueueItem) {    
    context.bindings.outputtable = {
        "partitionkey": "system",
        "status": myqueueitem.avgtemp
    context.log('Avg. temp: ', myQueueItem.avgtemp);

Surprise! Even if the function is written as in the book, an error will pop-up. This happens because in this moment Azure Function allows us only to add new rows to an Azure Table. Updating existing one is not possible.
The good news is that there is a small hack that we can do. Nothing block us to install Azure Storage module from npm and use it directly inside our function. To do this, you'll need to access the Kudu portal of you Azure Function. The URL is https://[FunctionAppName] and you'll authenticate with Azure credentials.
From there, go to 'Debug Console' and type 'npm install azure-storage'. Once you install this module you can make any calls to Azure Storage using Node.JS module.
process.env["AZURE_STORAGE_ACCOUNT"] = "vunvuleariotstorage";
process.env["AZURE_STORAGE_ACCESS_KEY"] = "@@@";
process.env["AZURE_STORAGE_CONNECTION_STRING "] = "@@@";

module.exports = function (context, myQueueItem) {
    var azure = require('azure-storage');
    var tableSvc = azure.createTableService();
    tableSvc.createTableIfNotExists('systemstatus', function (error, result, response) {
        if (!error) {
            var entityToUpdate = new Object();
            entityToUpdate.PartitionKey = "system";
            entityToUpdate.RowKey = "currenttemperature";
            entityToUpdate.Status = myQueueItem.avgtemp.toString();

            tableSvc.insertOrReplaceEntity('systemstatus', entityToUpdate, function (error, result, response) {
                if (!error) {
                    context.log('Avg. temp: ', myQueueItem.avgtemp);


I prefered to use process environments values to specify storage connection information. There is another way to do this, by specifying this data when you make the call to 'createTableService'. There is no error handling done, being a sample code.

ASP.NET Core that plays the role of heating system
There not to many things to say about it. It is a simple web application, that display minimum temperature, current temperature and heating system status.


  • There is not yet full support for all Azure Storage actions from Azure Function
  • You can load any module in Azure Function using NPM
  • You can add or remove inputs and outputs for of a Stream Analytics as you wish


Popular posts from this blog

ADO.NET provider with invariant name 'System.Data.SqlClient' could not be loaded

Today blog post will be started with the following error when running DB tests on the CI machine:
threw exception: System.InvalidOperationException: The Entity Framework provider type 'System.Data.Entity.SqlServer.SqlProviderServices, EntityFramework.SqlServer' registered in the application config file for the ADO.NET provider with invariant name 'System.Data.SqlClient' could not be loaded. Make sure that the assembly-qualified name is used and that the assembly is available to the running application. See for more information. at System.Data.Entity.Infrastructure.DependencyResolution.ProviderServicesFactory.GetInstance(String providerTypeName, String providerInvariantName) This error happened only on the Continuous Integration machine. On the devs machines, everything has fine. The classic problem – on my machine it’s working. The CI has the following configuration:

TeamCity.NET 4.51EF 6.0.2VS2013
It seems that there …

GET call of REST API that contains '/'-slash character in the value of a parameter

Let’s assume that we have the following scenario: I have a public HTTP endpoint and I need to post some content using GET command. One of the parameters contains special characters like “\” and “/”. If the endpoint is an ApiController than you may have problems if you encode the parameter using the http encoder.
using (var httpClient = new HttpClient()) { httpClient.BaseAddress = baseUrl; Task<HttpResponseMessage> response = httpClient.GetAsync(string.Format("api/foo/{0}", "qwert/qwerqwer"))); response.Wait(); response.Result.EnsureSuccessStatusCode(); } One possible solution would be to encode the query parameter using UrlTokenEncode method of HttpServerUtility class and GetBytes method ofUTF8. In this way you would get the array of bytes of the parameter and encode them as a url token.
The following code show to you how you could write the encode and decode methods.

Entity Framework (EF) TransactionScope vs Database.BeginTransaction

In today blog post we will talk a little about a new feature that is available on EF6+ related to Transactions.
Until now, when we had to use transaction we used ‘TransactionScope’. It works great and I would say that is something that is now in our blood.
using (var scope = new TransactionScope(TransactionScopeOption.Required)) { using (SqlConnection conn = new SqlConnection("...")) { conn.Open(); SqlCommand sqlCommand = new SqlCommand(); sqlCommand.Connection = conn; sqlCommand.CommandText = ... sqlCommand.ExecuteNonQuery(); ... } scope.Complete(); } Starting with EF6.0 we have a new way to work with transactions. The new approach is based on Database.BeginTransaction(), Database.Rollback(), Database.Commit(). Yes, no more TransactionScope.
In the followi…