In my series of blog posts about working with ARM templates I have gone through a lot of different use cases. In the last post I covered how to deploy a nested template to keep the main template a lot cleaner. One thing all posts have in common is that I´m using Azure Monitor as the main track which I´m deploying. I have deployed dozens of Log Analytics Workspaces for this. Now it´s time for deploying Azure Monitor alerts as well.Continue reading
In my last post which you can find here, I wrote about how to get started writing ARM templates for you Azure deployments. In the end I provided an ARM template built solely for that blog post. But the template was really basic and didn’t have much logic built into it, except for addressing a parameter to name the automation account and a variable to set the location where to create the automation account.Continue reading
Ever since Azure was a new thing and before it “went live”, we´ve been fed with how easy it is to get started using different services. One of my first experiences were when I deployed my blog on to the Azure platform somewhere back in 2012, and I was up and running in literally less than fifteen minutes, and this was the first time I did it which is why it took that long a time. It was just as simple as picking a WordPress instance from the gallery and deploying it. A similar experience was when I deployed my first virtual machine. A few clicks and a short time for the deployment and I had my new virtual machine ready to connect to.Continue reading
More and more services are moving to the cloud, Azure is one of the larges players but AWS and Google Cloud are also two large players. But just because resources are moving to the cloud doesn’t change the fact that we need to know how our environment is doing. Since I´m a monitoring guy, I write a lot about Azure Monitor and the capabilities of it to help us monitor our resources in the best possible way. But there is another aspect I want to touch as well, Azure service health.
While we monitor our resources using Azure Monitor, who monitors Azure Monitor as a service? Microsoft of course monitors all the Azure services to keep track of the status and to take immediate action when something goes down. We have the possibility to check up on Azure services from within Azure Monitor and that´s what I will be telling you more about with this post.Continue reading
The available Azure Monitor data sources is an interesting topic. Azure Monitor is a really powerful monitoring solution solely based in Azure, with a lot of capabilities. When the now retired Operations Management Suite were first presented, it was presented as a cloud agnostic solution meaning you could place your resources in any cloud besides from Azure, such as Amazon AWS or Google Cloud and still being able to monitor the resources. This is of course a real good thing (and necessary) since not everyone has or will have all their applications or servers in Azure only, there´s still a lot of on-prem servers and applications but also a lot of resources deployed in other public clouds as well.Continue reading
Azure Monitor is a quite new addition to the monitoring sphere when talking about monitoring Microsoft technologies. Traditionally it has been System Center Operations Manager (SCOM) that´s been the go-to guy but with the new addition of Azure Monitor some things have changed. From time to tome one can hear the phrase “SCOM is dead” and that you should go all-in with Azure Monitor instead. But is it really that easy?
In my opinion, no it’s not. While Azure Monitor has a lot of strengths being cloud-based with regularly updates and additions, it still lacks some things that we´re used to from using SCOM for all these years.
A while ago we were involved in a project with one of our customers where our goal was to connect a large amount of servers to Azure Log Analytics. They had already done this with a connection through SCOM, but when they added another management group to their servers, so called multi-homing it stopped working. This was a huge issue since the data they were sending to Log Analytics were really important to them, and we started discussing how to do this the best way. We eventually decided to go with a direct connection to Log Analytics, instead of going through SCOM as they had done before.Continue reading
Last week on January 15th was the last day of the OMS portal before its retirement. It has now completely moved to the Azure portal instead.
Since Operations Management Suite (OMS) have been retired for a few months and is no longer available for new customers, the portal had served its purpose and have now been retired. Nowadays administration of the included services is handled through the Azure portal instead.Continue reading
Lately I have been working a lot with monitoring VMware using SCOM for some of our largest customers and have gotten to think about this more and more. Even though cloud providers such as Amazon Web Services (AWS) and Microsoft Azure keeps on showing great numbers of growth (and profits for that matter), the absolute majority of customers IT are still on-prem. Since about ten years, virtualization has been about the coolest thing there is and the largest player in this area is still VMware.
Thinking about how large this area is and the importance for the organization, we need to monitor the VMware platform. Just as well as we need to keep track of what’s happening with our services, such as web shops or other business critical systems, we need to monitor the foundation it all relies on as well.
Recently Altaro VM Backup was released in a new version, 7.6. This VM backup solution works for Hyper-V and VMware and this version introduces a bunch of new features. In my lab environment I´m using their backup solution to back up my servers. I have written about this once before a couple of years back but since a lot has happened during the years, it´s time to write a new review.
The new stuff in this version are the following;
Continuous Data Protection (CDP)
When CDP is enabled, VMs can be backed up as frequently as every 5 minutes or the maximum frequency permitted by the backup location and environment. This ensures that in a data loss scenario only a few minutes of data would be lost.
Grandfather-Father-Son (GFS) Archiving
GFS enables users to choose to archive the backup versions over and above their continuous and daily backups instead of deleting them (local backups only).
Now you can easily set up separate backup cycles to store a new backup version every week, every month and every year.
And then a couple of bug fixes. Read more about the news and bug fixes here.