Welcome!

@DevOpsSummit Authors: Kelly Burford, Automic Blog, Derek Weeks, Liz McMillan, Elizabeth White

Related Topics: @DevOpsSummit, Microsoft Cloud, Open Source Cloud, @CloudExpo

@DevOpsSummit: Article

Windows Azure Resource Manager (ARM) | @CloudExpo @Azure #Cloud #IoT

An overview ARM and the new features it provides

Over the past year, Microsoft has been introducing Azure Resource Manager (ARM) as the preferred way to provision and manage resources in its Azure cloud. ARM is the successor to the original Service Management model, also known as "Classic."[1] While Classic will continue to be supported for the foreseeable future, ARM is the preferred deployment model and all new Azure features are being released on ARM only. Here is an overview ARM and the new features it provides.

ARM introduces the concept of "Resource Groups" and "Templates." Resource Groups are a simple way to manage and group related resources. In addition to simplifying the management of related resources, they also simplify accounting, security and auditing of your Azure deployment. Resource Manager Templates give Azure administrators a very powerful tool that they can leverage to rapidly deploy complex environments consisting of interrelated resources rapidly and consistently. Although there is a learning curve to effectively leverage Resource Manager Templates, once mastered this will be the most effective means of deploying resources in Azure.

Resource Manager Templates
While Resource Manager Templates are very powerful, administrators can still provision new resources in Azure using tools that they are familiar with. In addition to the Azure Portal GUI, Microsoft has released Azure PowerShell 1.0.[2] PowerShell has been the most efficient tool for managing Azure Classic and will continue to work just as well in ARM. For those administrators managing Azure from Mac OS or Linux there is also Azure Command Line which allows Azure management from a set of open-source shell-based commands.[3]

While the new Azure Portal has a variety of improvements over the previous portal and PowerShell continues to be a powerful tool for provisioning and managing resources, the real power of ARM comes from Resource Manager Templates, a declarative way to define your deployments.[4]

Resource Manager Templates
Resource Manager Templates are written in JavaScript Object Notation (JSON) and are used to define your resources as well as to create dependencies between resources. Microsoft has published a number of Azure Quickstart Templates which can be used "as is" or as examples as you build your own templates.[5] Templates generally allow users to pass parameters so each deployment can be customized, to an extent, at deployment time.

Templates can also be linked together so that a single main template can call multiple linked templates. These linked templates can pass parameters back and forth from the master template to the linked templates. This allows you to write templates that perform specific functions and reuse them in other deployments, much as a programmer would define functions within an application for reuse.

While this new paradigm is powerful, it poses a challenge for some Windows administrators who are concerned about programming languages. Many administrators are just now becoming comfortable with PowerShell; others prefer to use the Windows GUI whenever possible. While JSON is a relatively simple language to read and write, many Windows administrators will avoid using Templates because of their lack of experience with JSON and with programming languages in general.

Windows administrators who find themselves suddenly charged with managing Windows Azure deployments can still use a combination of the Azure Portal and Azure PowerShell commands. However, you cannot just use the Azure Portal UI because some ARM features are currently only available via Azure PowerShell. Azure Administrators should buy Visual Studio and learn how to write Resource Manager Templates in JSON.

Visual Studio has added several Azure-specific SDKs to expand Visual Studio. There are tools within Visual Studio that help developers build new programs and mobile applications that are cloud integrated. Windows administrators can also leverage Visual Studio to define and launch VMs and supporting resources right from within Visual Studio.[6]

Resource Groups
Azure Classic had a concept of a "Cloud Service", which was their attempt to group related resources. The limited functionality of Cloud Service has been replaced in ARM by Resource Groups which are used to manage the entire lifecycle of an application, from creation to deletion.

Resource Groups
Resource Groups are a container for all related Azure resources, from VMs, networks, IP addresses, load balancers, databases, web apps and more.  When creating new resources, you can choose to add a resource to an existing Resource Group or to create a new Resource Group. There are no explicit rules governing what should be placed in the same Resource Group, but there are a few things to keep in mind when making that decision.

Resource Groups are a security boundary. You can grant permissions at the Resource Group layer to different users or groups. Keep in mind, they are not NTFS security within your Windows Domain, rather, they provide security in terms of who has access to manage and/or monitor your Azure resources. At last count there were 32 different pre-defined security roles ranging from basic roles like Owner, Contributor and Reader all the way down to specific roles like DNS Zone Contributor and Security Manager.

When defining Resource Groups consider grouping resources that have similar security needs in the same Resource Group. Although permissions can be granted at the Resource Group layer, if there are sensitive VMs within a Resource Group that need additional security these VMs can have explicit security which overrides the security of the Resource Group, so you don't have to worry if your security needs are a little "complicated".

In addition to security, resource utilization, alerting, monitoring and billing can be done at the Resource  Group level as shown in Figure 1.

Figure 1: Example of a Resource Group panel in the Azure Portal

The Resource Group panel displays a simple overview of all the resources in a Resource Group and allows you directly click on each resource to drill down to get a more detailed view of resource as well as allow you to make changes to those resources. This panel is customizable, allowing you to display additional detail such as estimated spend, disk, CPU and memory utilization as well.

If you have ever managed an Azure Classic deployment, or even fooled around with Azure Classic in your home lab you will appreciate this next feature. When it comes time to decommission your application, the old method was to delete you your resources, one resource at a time. And don't forget; before you delete that Storage Account be sure to delete all the VHD files, all the containers and all the blobs first!  It wasn't pretty.

Cleaning up a failed deployment or simply trying to reset your lab to a clean slate in Azure Classic could sometimes be a tedious process. Forgetting to delete just one Premium Storage Disk after a POC will come as quite a surprise at the end of the month when your receive your credit card statement full of unexpected Azure charges; don't ask me how I know.

ARM simplifies this whole process. When you are done with an application and want to decommission it completely, just click "Delete" and all the resources within the Resource Group will be deleted, no fuss no muss (See Figure 2). Now as you can imagine, this is a pretty destructive command, so they do ask you to confirm what you are about to do and ask you to type the name of the Resource Group once again before the job will run to ensure you really are deleting the Resource Group you intended.

Figure 2: Deleting Resource Groups

Summary
In addition to Resource Groups and Template deployments there are other new features of ARM, such as the ability to perform tasks in parallel rather than series. For example, in Classic to stop four VMs via the portal, you would have to stop the first VM, wait for a while, and then try to stop the second VM. If the first VM was not far enough along in the stopping process my attempt to stop the second VM would fail. Needless to say you only had to do this a few times before you decided to write a PowerShell script to stop all your VMs.

With ARM you can now perform this same task in parallel, you no longer have to wait for the first VM to stop before you can stop the second, third and fourth VM. The same holds true for other operations that were previously serialized, such as creating new VMs.

While ARM is the future of Azure some features took a long time to be added and some are not as well documented as they are in Azure Classic. However, the benefits described here should help you overcome any hesitation you may have about making ARM your preferred Azure deployment method.

Learn more about ARM in this recorded webinar.

References

  1. https://azure.microsoft.com/en-us/documentation/articles/resource-manager-deployment-model/
  2. https://azure.microsoft.com/en-us/blog/azps-1-0/
  3. https://azure.microsoft.com/en-us/documentation/articles/xplat-cli-install/
  4. https://azure.microsoft.com/en-us/documentation/articles/resource-group-authoring-templates/
  5. https://azure.microsoft.com/en-us/documentation/templates/
  6. https://www.visualstudio.com/en-us/features/azure-tools-vs.aspx

More Stories By David Bermingham

David Bermingham is recognized within the technology community as a high availability expert and has been honored by his peers by being elected to be a Microsoft MVP in Clustering since 2010. His work as director of Technical Evangelist at SIOS has him focused on evangelizing Microsoft high availability and disaster recovery solutions as well as providing hands on support, training and professional services for cluster implementations.

David holds numerous technical certifications and draws from over twenty years of experience in IT, including work in the finance, healthcare and education fields, to help organizations design solutions to meet their high availability and disaster recovery needs. He has recently begun speaking on deploying highly available SQL Servers in the Azure Cloud and deploying Azure Hybrid Cloud for disaster recovery.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@DevOpsSummit Stories
In his Opening Keynote at 21st Cloud Expo, John Considine, General Manager of IBM Cloud Infrastructure, led attendees through the exciting evolution of the cloud. He looked at this major disruption from the perspective of technology, business models, and what this means for enterprises of all sizes. John Considine is General Manager of Cloud Infrastructure Services at IBM. In that role he is responsible for leading IBM’s public cloud infrastructure including strategy, development, and offering management. To date, IBM has launched more than 50 cloud data centers that span the globe. He has been building advanced technology, delivering “as a service” solutions, and managing infrastructure services for the past 20 years.
The past few years have brought a sea change in the way applications are architected, developed, and consumed—increasing both the complexity of testing and the business impact of software failures. How can software testing professionals keep pace with modern application delivery, given the trends that impact both architectures (cloud, microservices, and APIs) and processes (DevOps, agile, and continuous delivery)? This is where continuous testing comes in. D
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the benefits of the cloud without losing performance as containers become the new paradigm.
Digital transformation is about embracing digital technologies into a company's culture to better connect with its customers, automate processes, create better tools, enter new markets, etc. Such a transformation requires continuous orchestration across teams and an environment based on open collaboration and daily experiments. In his session at 21st Cloud Expo, Alex Casalboni, Technical (Cloud) Evangelist at Cloud Academy, explored and discussed the most urgent unsolved challenges to achieve full cloud literacy in the enterprise world.
The 22nd International Cloud Expo | 1st DXWorld Expo has announced that its Call for Papers is open. Cloud Expo | DXWorld Expo, to be held June 5-7, 2018, at the Javits Center in New York, NY, brings together Cloud Computing, Digital Transformation, Big Data, Internet of Things, DevOps, Machine Learning and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal today!
In a recent survey, Sumo Logic surveyed 1,500 customers who employ cloud services such as Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP). According to the survey, a quarter of the respondents have already deployed Docker containers and nearly as many (23 percent) are employing the AWS Lambda serverless computing framework. It’s clear: serverless is here to stay. The adoption does come with some needed changes, within both application development and operations. That means serverless is also changing the way we leverage public clouds. Truth-be-told, many enterprise IT shops were so happy to get out of the management of physical servers within a data center that many limitations of the existing public IaaS clouds were forgiven. However, now that we’ve lived a few years with public IaaS clouds, developers and CloudOps pros are giving a huge thumbs down to the ...
Kubernetes is an open source system for automating deployment, scaling, and management of containerized applications. Kubernetes was originally built by Google, leveraging years of experience with managing container workloads, and is now a Cloud Native Compute Foundation (CNCF) project. Kubernetes has been widely adopted by the community, supported on all major public and private cloud providers, and is gaining rapid adoption in enterprises. However, Kubernetes may seem intimidating and complex to learn. This is because Kubernetes is more of a toolset than a ready solution. Hence it’s essential to know when and how to apply the appropriate Kubernetes constructs.
22nd International Cloud Expo, taking place June 5-7, 2018, at the Javits Center in New York City, NY, and co-located with the 1st DXWorld Expo will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterprises are using some form of XaaS – software, platform, and infrastructure as a service.
22nd International Cloud Expo, taking place June 5-7, 2018, at the Javits Center in New York City, NY, and co-located with the 1st DXWorld Expo will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterprises are using some form of XaaS – software, platform, and infrastructure as a service.
DevOps at Cloud Expo – being held June 5-7, 2018, at the Javits Center in New York, NY – announces that its Call for Papers is open. Born out of proven success in agile development, cloud computing, and process automation, DevOps is a macro trend you cannot afford to miss. From showcase success stories from early adopters and web-scale businesses, DevOps is expanding to organizations of all sizes, including the world's largest enterprises – and delivering real results. Among the proven benefits, DevOps is correlated with 20% faster time-to-market, 22% improvement in quality, and 18% reduction in dev and ops costs, according to research firm Vanson-Bourne. It is changing the way IT works, how businesses interact with customers, and how organizations are buying, building, and delivering software.
All clouds are not equal. To succeed in a DevOps context, organizations should plan to develop/deploy apps across a choice of on-premise and public clouds simultaneously depending on the business needs. This is where the concept of the Lean Cloud comes in - resting on the idea that you often need to relocate your app modules over their life cycles for both innovation and operational efficiency in the cloud.
@DevOpsSummit at Cloud Expo, taking place June 5-7, 2018, at the Javits Center in New York City, NY, is co-located with 22nd Cloud Expo | 1st DXWorld Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete at launch. DevOps may be disruptive, but it is essential.
Cloud Expo | DXWorld Expo have announced the conference tracks for Cloud Expo 2018. Cloud Expo will be held June 5-7, 2018, at the Javits Center in New York City, and November 6-8, 2018, at the Santa Clara Convention Center, Santa Clara, CA. Digital Transformation (DX) is a major focus with the introduction of DX Expo within the program. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term. A total of 88% of Fortune 500 companies from a generation ago are now out of business. Only 12% still survive. Similar percentages are found throughout enterprises of all sizes.
SYS-CON Events announced today that T-Mobile exhibited at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. As America's Un-carrier, T-Mobile US, Inc., is redefining the way consumers and businesses buy wireless services through leading product and service innovation. The Company's advanced nationwide 4G LTE network delivers outstanding wireless experiences to 67.4 million customers who are unwilling to compromise on quality and value. Based in Bellevue, Washington, T-Mobile US provides services through its subsidiaries and operates its flagship brands, T-Mobile and MetroPCS. For more information, visit https://www.t-mobile.com.
SYS-CON Events announced today that Cedexis will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 - Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Cedexis is the leader in data-driven enterprise global traffic management. Whether optimizing traffic through datacenters, clouds, CDNs, or any combination, Cedexis solutions drive quality and cost-effectiveness. For more information, please visit https://www.cedexis.com.
SYS-CON Events announced today that Google Cloud has been named “Keynote Sponsor” of SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Companies come to Google Cloud to transform their businesses. Google Cloud’s comprehensive portfolio – from infrastructure to apps to devices – helps enterprises innovate faster, scale smarter, stay secure, and do more with data than ever before.
Gemini is Yahoo’s native and search advertising platform. To ensure the quality of a complex distributed system that spans multiple products and components and across various desktop websites and mobile app and web experiences – both Yahoo owned and operated and third-party syndication (supply), with complex interaction with more than a billion users and numerous advertisers globally (demand) – it becomes imperative to automate a set of end-to-end tests 24x7 to detect bugs and regression. In their session at 21st Cloud Expo, Jenny Hung, E2E Engineer Manager at Yahoo Gemini, Haoran Zhao, Software Engineer at Oath Gemini, and Lin Zhang, Software Engineer at Oath (Yahoo), will describe the technical challenges and the principles we followed to build a reliable and scalable test automation infrastructure across desktops, mobile apps, and mobile web platforms on the cloud. We also share some...
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm. In their Day 3 Keynote at 20th Cloud Expo, Chris Brown, a Solutions Marketing Manager at Nutanix, and Mark Lavi, a Nutanix DevOps Solution Architect, explored the ways that Nutanix technologies empower teams to react faster than ever before and connect teams in ways that were either too complex or simply impossible with traditional infrastructures.
SYS-CON Events announced today that Vivint to exhibit at SYS-CON's 21st Cloud Expo, which will take place on October 31 through November 2nd 2017 at the Santa Clara Convention Center in Santa Clara, California. As a leading smart home technology provider, Vivint offers home security, energy management, home automation, local cloud storage, and high-speed Internet solutions to more than one million customers throughout the United States and Canada. The end result is a smart home solution that saves you time and money and ultimately simplifies your life.
SYS-CON Events announced today that Opsani will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Opsani is the leading provider of deployment automation systems for running and scaling traditional enterprise applications on container infrastructure.
Every few years, a disruptive force comes along that prompts us to reframe our understanding of what something means, or how it works. For years, the notion of what a computer is and how you make one went pretty much unchallenged. Then virtualization came along, followed by cloud computing, and most recently containers. Suddenly the old rules no longer seemed to apply, or at least they didn’t always apply. These disruptors made us reconsider our IT worldview.
SYS-CON Events announced today that Nirmata will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Nirmata provides a comprehensive platform, for deploying, operating, and optimizing containerized applications across clouds, powered by Kubernetes. Nirmata empowers enterprise DevOps teams by fully automating the complex operations and management of application containers and its underlying resources. Nirmata not only simplifies deployment and management of Kubernetes clusters but also facilitates delivery and operations of applications by continuously monitoring the application and infrastructure for changes, and auto-tuning the application based on pre-defined policies. Using Nirmata, enterprises can accelerate their journey towards becoming cloud-native.
SYS-CON Events announced today that Opsani to exhibit at SYS-CON's 21st Cloud Expo, which will take place on October 31 through November 2nd 2017 at the Santa Clara Convention Center in Santa Clara, California. Opsani is creating the next generation of automated continuous deployment tools designed specifically for containers. How is continuous deployment different from continuous integration and continuous delivery? CI/CD tools provide build and test. Continuous Deployment is the means by which qualified changes in software code or architecture are automatically deployed to production as soon as they are ready. Adding continuous deployment to your toolchain is the final step to providing push button deployment for your developers.
Coca-Cola’s Google powered digital signage system lays the groundwork for a more valuable connection between Coke and its customers. Digital signs pair software with high-resolution displays so that a message can be changed instantly based on what the operator wants to communicate or sell. In their Day 3 Keynote at 21st Cloud Expo, Greg Chambers, Global Group Director, Digital Innovation, Coca-Cola, and Vidya Nagarajan, a Senior Product Manager at Google, will discuss how from store operations and optimization to employee training and insights, all ultimately create the best customer experience both online and in-store.
The next XaaS is CICDaaS. Why? Because CICD saves developers a huge amount of time. CD is an especially great option for projects that require multiple and frequent contributions to be integrated. But… securing CICD best practices is an emerging, essential, yet little understood practice for DevOps teams and their Cloud Service Providers. The only way to get CICD to work in a highly secure environment takes collaboration, patience and persistence. Building CICD in the cloud requires rigorous architectural and coordination work to minimize the volatility of the cloud environment and leverage the security features of the cloud to the benefit of the CICD pipeline.