Skip to main content

[IoT Home Project] Part 5 - Send data to Azure IoT Hub, control time interval and refac the configuration information

In this post we will discover how we can:
  • Send all device sensor data that are read from GrovePI to Azure IoT Hub
  • Read sensor data at a specific time interval
  • Extract all configuration data in a separate file

Send all device sensor data that are read from GrovePI to Azure IoT Hub
This is a simple task. The function that reads sensor data from GrovePI already returns all the sensor data. The only thing that we need to do is to put this information in the message that is send to Azure IoT Hub.
In the future it is pretty clear that we will have different type of messages that we send over IoT Hub. Because of this we shall add a property to the message that is send to IoT Hub that specifies the message type - in our case we'll call the message that contains sensor data 'sensorData'.
var dataToSend = JSON.stringify({
    deviceId: "vunvulearaspberry",
    msgType: "sensorData",
    sensorInf: {
        temp: sensorsData.temp,
        humidity: sensorsData.humidity,
        distance: sensorsData.distance,
        light: sensorsData.light
    }
});
deviceCommunication.sendMessage(dataToSend);

Read sensor data at a specific time interval
In this moment we have a 'while(true)' that reads sensor information and send data to IoT Hub. This works great, but what if we want to control how often data is read.
To be able to do this, we can use 'setInterval' function from node.js. This function allows us to specify a time interval when a function is called. The time interval is specified in milliseconds. 
The nice thing is that the functions will not be executed in parallel if the first call didn't finished yet. This is important when we specify a small time interval like 100 ms and the read sensor information and send data operations takes more than 100 ms.
function collectSensorData(grovePiSensors, deviceCommunication) {
    var timeIntervalInMilisec = 5000;  // 5s
    setInterval((grovePiSensors, deviceCommunication) => {
        var sensorsData = grovePiSensors.getAllSensorsData();

        var dataToSend = JSON.stringify({
            deviceId: "vunvulearaspberry",
            msgType: "sensorData",
            sensorInf: {
                temp: sensorsData.temp,
                humidity: sensorsData.humidity,
                distance: sensorsData.distance,
                light: sensorsData.light
            }

        });
        deviceCommunication.sendMessage(dataToSend);
    }, timeIntervalInMilisec, grovePiSensors, deviceCommunication);
}

Extract all configuration data in a separate file
In this moment we have configuration data in multiple modules. If we want to change something, we need to search where the configuration is stored, some information like device id string is duplicated.
The code is not so nice and a change can be buggy and time consuming.
To avoid all this problems, we ca create a config.json file in our application where all configuration data is added. For each module I prefer to create a section where I group specific configuration for that that module ('deviceCommunicationConfig' and 'grovePiConfig').
{
    "debug" : true,
    "sensorDataTimeSampleInSec" : 5,
    "deviceCommunicationConfig":
    {
        "deviceId" : "vunvulearaspberry",
        "azureIoTHubMasterConnectionString" : "HostName=vunvulear-iot-hub.azure-devices.net;SharedAccessKeyName=iothubowner;SharedAccessKey=+whKMyd08PLDNoaR+yEmToJcHL6wsFo36tAyDBU8Qr0=",
        "azureIoTHubHostName" : "vunvulear-iot-hub.azure-devices.net"
    },
    "grovePiConfig":
    {
        "dhtDigitalSensorPin" : 2,
        "ultrasonicDigitalSensorPin" : 4,
        "lightAnalogSensorPin" : 2,
        "soundAnalogSensorPin" : 0
    }
}

Once we have all this configuration, we will need to access it. For this we can load the JSON in our application and access the properties of the configuration file.
var Config = require('./config.json');
...
var grovePiSensors = new GrovePiSensors(Config.grovePiConfig);
...
this.debug = debug;
this.registry = AzureIoTHub.Registry.fromConnectionString(Config.deviceCommunicationConfig.azureIoTHubMasterConnectionString);
this.deviceId = Config.deviceCommunicationConfig.deviceId;
this.azureIoTHubHostName = Config.deviceCommunicationConfig.azureIoTHubHostName;

In each module we don't need to specify all the configuration file. We can only specify the section that is specific to that module ('Config.deviceCommunicationConfig'). For this you can take a look in the source file.

Conclusion
Mission complete for now. See push all sensor data to Azure IoT Hub and we have the flexibility to change configuration more easily. Additional to this we have control on how often data is pushed to Azure IoT Hub.

Next Step
In the next post we will start to store the data that is pushed to Azure IoT Hub to blobs. On top of this we will take a look on how we can calculate the average values for read data.

Next post: [IoT Home Project] Part 6 - Stream Analytics and Power BI

Comments

Popular posts from this blog

Windows Docker Containers can make WIN32 API calls, use COM and ASP.NET WebForms

After the last post , I received two interesting questions related to Docker and Windows. People were interested if we do Win32 API calls from a Docker container and if there is support for COM. WIN32 Support To test calls to WIN32 API, let’s try to populate SYSTEM_INFO class. [StructLayout(LayoutKind.Sequential)] public struct SYSTEM_INFO { public uint dwOemId; public uint dwPageSize; public uint lpMinimumApplicationAddress; public uint lpMaximumApplicationAddress; public uint dwActiveProcessorMask; public uint dwNumberOfProcessors; public uint dwProcessorType; public uint dwAllocationGranularity; public uint dwProcessorLevel; public uint dwProcessorRevision; } ... [DllImport("kernel32")] static extern void GetSystemInfo(ref SYSTEM_INFO pSI); ... SYSTEM_INFO pSI = new SYSTEM_INFO(...

How to audit an Azure Cosmos DB

In this post, we will talk about how we can audit an Azure Cosmos DB database. Before jumping into the problem let us define the business requirement: As an Administrator I want to be able to audit all changes that were done to specific collection inside my Azure Cosmos DB. The requirement is simple, but can be a little tricky to implement fully. First of all when you are using Azure Cosmos DB or any other storage solution there are 99% odds that you’ll have more than one system that writes data to it. This means that you have or not have control on the systems that are doing any create/update/delete operations. Solution 1: Diagnostic Logs Cosmos DB allows us activate diagnostics logs and stream the output a storage account for achieving to other systems like Event Hub or Log Analytics. This would allow us to have information related to who, when, what, response code and how the access operation to our Cosmos DB was done. Beside this there is a field that specifies what was th...

Cloud Myths: Cloud is Cheaper (Pill 1 of 5 / Cloud Pills)

Cloud Myths: Cloud is Cheaper (Pill 1 of 5 / Cloud Pills) The idea that moving to the cloud reduces the costs is a common misconception. The cloud infrastructure provides flexibility, scalability, and better CAPEX, but it does not guarantee lower costs without proper optimisation and management of the cloud services and infrastructure. Idle and unused resources, overprovisioning, oversize databases, and unnecessary data transfer can increase running costs. The regional pricing mode, multi-cloud complexity, and cost variety add extra complexity to the cost function. Cloud adoption without a cost governance strategy can result in unexpected expenses. Improper usage, combined with a pay-as-you-go model, can result in a nightmare for business stakeholders who cannot track and manage the monthly costs. Cloud-native services such as AI services, managed databases, and analytics platforms are powerful, provide out-of-the-shelve capabilities, and increase business agility and innovation. H...