Interactive analysis have become a major part of the field of Data Science. Two tools have become very popular, Jupyter and Zeppelin.
This article will show you how to provision a Spark cluster and run analysis on it with the help of Zeppelin.
The Internet of Things (IoT) has finally arrived (some time ago). An important part is to analyze sensor data in motion. For that a streaming system is necessary. You could use Apache Storm or Apache Spark Streaming to do that, but if you want to run it as a service on Azure without going through the pain to set up a cluster Azure Stream Analytics is a good choice.
In this post I’m going through the basics of Azure Stream Analytics and DocumentDB, which is used as a destination for our data after the streaming is done, and how you can create a simple Stream Analytics job using Blob storage as the input and DocumentDB as the output.
Continous delivery (or continuous deployment or continuous integration) is an important part of the modern software development lifecycle.
Azure WebJobs on the other hand is a nice little feature to run processes in the background.
This article describes in short why continuous delivery is a good thing, what Azure WebJobs are and how you can use both together.
It’s getting cold outside and November already arrived and there are just a few days left until Christmas. This means that it is time for another Links of the month.
This issue contains some interesting blog post related to Azure and Apache Spark in general.