Welcome!

@DevOpsSummit Authors: Stefana Muller, Karthick Viswanathan, Yeshim Deniz, Liz McMillan, Zakia Bouachraoui

Related Topics: @CloudExpo, Java IoT, Linux Containers, @DXWorldExpo, SDN Journal, @DevOpsSummit

@CloudExpo: Blog Post

Monitoring Magic and the Future of APM

The Cloud Precipitating New Agents for Monitoring

At the time when we were looking for a monitoring solution (2006-2007) APM as we know it today had yet to be defined.

There was no Gartner MQ, real-user-monitoring (RUM) was too high level, “agent monitoring” brought concerns of overhead and complexity, instrumenting the application meant to ARM it (i.e., Application Response Measurement), and transaction tagging was a pipe dream.

This created a fierce debate on the risks and rewards of agent vs. agentless monitoring, read The Monitoring Duality of APM. So, when we were developing our monitoring approach, our first priority was to do no harm, then collect performance metrics.

We first implemented agentless RUM technology (i.e. wire data analytics) to gain insight into the application behavior and build a baseline that captured a normal workload. Secondly we focused on synthetic transactions to provide visibility during off peak hours when transaction volumes were low.

Months later, after we had a basic end-user-experience (EUE) solidified and earned the trust of the stakeholders, we began working on deploying agent-based monitoring in production to get a runtime view of the J2EE stack tying it back to the business transactions.

Today’s APM has come a long way since then and has advanced to levels that some may refer to as “magic.” But with these new found powers, APM now comes with a paparazzi of peddlers trying to sell you something, and hucksters who are preaching how to reach the promised land of performance through marketing glossies.

Deciding on an APM solution that's right for your environment can be a daunting task with so many vendors using APM in their messaging today. It would be wise to start small, curate a POC, and get comfortable with the solution nuances, scrutinizing the integration touch points that will coalesce with your current operational processes.

Since there are so many facets to consider when selecting a modern-day APM solution, be wary of the “one size fits all” approach no matter how compelling the story. Remember, you are striving for a robust, real-time application monitoring solution which will most likely utilize a combination of active and passive monitoring techniques.

Consider that the Age of the Cloud has precipitated a new agent concept which is lightweight, deploys quickly, and goes in virtually undetected with zero configuration. These agents are built with a survival mode in mind including a self-healing option for hands-free maintenance.

Looking into the future, I see an evolution towards an intelligent transparent agent, that once given a “target application” to monitor, will run autonomously, and spawn automatically as virtual resources shift. The inherent behavioral learning capabilities will allow it to self-regulate and adjust system resources accordingly, utilizing adaptive controls to support predefined SLA's.

Continue reading, click here for the full article.

Image: Snugg LePup/Flickr

More Stories By Larry Dragich

Larry Dragich is actively involved with industry leaders, sharing knowledge of Application Performance Management (APM) technologies, from best practices and technical workflows, to resource allocation and approaches for implementation. He has been working in the APM space since 2006 where he built the Enterprise Systems Management team which is now the focal point for IT performance monitoring and capacity planning activities.

@DevOpsSummit Stories
As Cybric's Chief Technology Officer, Mike D. Kail is responsible for the strategic vision and technical direction of the platform. Prior to founding Cybric, Mike was Yahoo's CIO and SVP of Infrastructure, where he led the IT and Data Center functions for the company. He has more than 24 years of IT Operations experience with a focus on highly-scalable architectures.
An edge gateway is an essential piece of infrastructure for large scale cloud-based services. In his session at 17th Cloud Expo, Mikey Cohen, Manager, Edge Gateway at Netflix, detailed the purpose, benefits and use cases for an edge gateway to provide security, traffic management and cloud cross region resiliency. He discussed how a gateway can be used to enhance continuous deployment and help testing of new service versions and get service insights and more. Philosophical and architectural approaches to what belongs in a gateway vs what should be in services were also discussed. Real examples of how gateway services are used in front of nearly all of Netflix's consumer facing traffic showed how gateway infrastructure is used in real highly available, massive scale services.
DXWorldEXPO LLC announced today that Telecom Reseller has been named "Media Sponsor" of CloudEXPO | DXWorldEXPO 2018 New York, which will take place on November 11-13, 2018 in New York City, NY. Telecom Reseller reports on Unified Communications, UCaaS, BPaaS for enterprise and SMBs. They report extensively on both customer premises based solutions such as IP-PBX as well as cloud based and hosted platforms.
Enterprises have taken advantage of IoT to achieve important revenue and cost advantages. What is less apparent is how incumbent enterprises operating at scale have, following success with IoT, built analytic, operations management and software development capabilities - ranging from autonomous vehicles to manageable robotics installations. They have embraced these capabilities as if they were Silicon Valley startups.
Wooed by the promise of faster innovation, lower TCO, and greater agility, businesses of every shape and size have embraced the cloud at every layer of the IT stack – from apps to file sharing to infrastructure. The typical organization currently uses more than a dozen sanctioned cloud apps and will shift more than half of all workloads to the cloud by 2018. Such cloud investments have delivered measurable benefits. But they’ve also resulted in some unintended side-effects: complexity and risk. End users now struggle to navigate multiple environments with varying degrees of performance. Companies are unclear on the security of their data and network access. And IT squads are overwhelmed trying to monitor and manage it all.