Welcome!

@DevOpsSummit Authors: Elizabeth White, Liz McMillan, Don MacVittie, Dalibor Siroky, Automic Blog

Related Topics: Microservices Expo, @CloudExpo, @DevOpsSummit

Microservices Expo: Article

Reference Architecture #Microservices | @CloudExpo #Serverless #AI #DevOps

The goal of microservices is to improve software delivery speed and increase system safety as scale increases

The goal of Microservices is to improve software delivery speed and increase system safety as scale increases. Microservices being modular these are faster to change and enables an evolutionary architecture where systems can change, as the business needs change. Microservices can scale elastically and by being service oriented can enable APIs natively. Microservices also reduce implementation and release cycle time and enables continuous delivery. This paper provides a logical overview of the Microservices Reference Architecture that highlights various sub systems needed to support Microservices deployment and execution.

Introduction
Switch over to Microservices is an hour of need in web application development and delivery and is crucial for the success of Enterprises today.

Of late, enterprises are adopting technologies like analytics, mobility, social media, IoT and smart embedded devices to change customer relationships, internal processes and value propositions. Microservices acts like a bridge between these technologies and provides the building blocks for developing modern distributed enterprise systems to become one of the enablers in the digital transformation journey of the enterprise.

Adoption of Microservices provides the agility, reliability, maintainability, scalability and deploy ability to the enterprises as part of digital transformation process.

There exists many different architecture definition approaches for implementing Microservices across Industry. Many of them are unique and specific to the needs of individual development teams. Many enterprises, without understanding or highlighting the name, have been using an approach towards leveraging APIs that classified as Microservices. This demands a need for designing Reference Architecture to develop and deliver Microservices‑based applications that consumed across the Enterprises.

There exists, numerous articles and blogs on Microservices reference architecture, architecture principles and best practices. This paper summarizes the purpose of Microservices Reference Architecture, drivers for adopting Microservices, Microservices architecture principles and logical view of Microservices Reference Architecture.

Drivers for Microservice Reference Architecture
Microservices are smaller in scope, determined by a focus on domain boundaries and consistent domain modelling, and require less code. In addition, in Monolithic applications, the application components communicate in memory. In the case of Microservices based applications, the communication happens over the network. This leads to software development and deployment becomes more reliable and faster.

With Microservices, each service can scale independently to meet temporary traffic, complete batch processing and other business needs. Improved fault isolation restricts service issues, such as memory leaks or open database connections. The scalability of Microservices complements the flexibility of cloud services improves service and handle more customers simultaneously without interrupting service.

The following diagram represents few key drivers for Enterprises adopting Microservices.

Microservices Architecture Principles
Below are few key architectural principles for Microservices:

Single Responsibility Principle: Each microservice must be responsible for a specific feature or a business functionality or aggregation of cohesive functionality completely.

Granular: Microservice granularity is contained within the intersection of a single functional domain, a single data domain and its immediate dependencies, a self-sufficient packaging and a technology domain.

Domain Driven Design: Domain driven design is an architectural principle in-line with object-oriented approach. It considers the business domain, elements and behaviors and interactions between business domains.

Encapsulation: Each Microservice encapsulates the internal implementation details, so that the external system utilizes the services need not worry about the internals. Encapsulation reduces the complexity and enhances the flexibility of the system.

Loose Coupling: There must be zero coordination necessary for the deployment with other Microservices. The changes in one microsystem should have zero or minimum impact on other services in the eco-system.  

Separation of Concerns: Develop the microservices based on distinct features with zero overlap with other functions. The main objective is to reduce the interaction between services so that they are highly cohesive and loosely coupled. If we separate, the functionality across wrong boundaries will lead tight coupling and increased complexity between services.

Language Neutral: Microservices are composed together to form a complex application, and they do not need to be written using the same programming language. For example, Java might be the correct language for one application, and in others, it might be Python etc.

Hexagonal Architecture: Microservice exposes RESTful APIs for external communication, message broker interface for event notification and database adapters for persistence. This makes hexagonal architecture as a most suitable style for Microservice development.

Sizing of Microservices
It is important that, while designing a Microservice, there is a need to decide on the number and size of individual Microservices. There is no strict rule regarding the optimal size of the Microservice and it depends on the partitioning of the problem space if it is a new development application or splitting the existing monolithic application into individual Microservices.

Also, Microservices cannot be too large or too small. Large services are hard to work with, hard to deploy, and take longer to start and stop. On the other side, when Microservices are too small, the resource cost of deploying and operating such a service overshadows its utility.

Microservice granularity can also be determined based on business needs. Making services too granular, or requiring too many dependencies on other Microservices, can introduce latency.

Microservices allow teams to plan, develop, and deploy features of a system in the cloud without tight coordination. It is therefore, Microservice number and size be dictated by business and technical principles.

Microservices Reference Architecture
Microservices is an architecture style, in which software systems or applications are composed of one or more independent and self-contained services. It is not a product, framework, or platform. It is a strategy for building large distributed systems and are loosely coupled and deployed independently of one another.

The following guidelines to be adapted while designing the services of Microservices Reference Architecture.

  • Lightweight: To facilitate smaller memory footprints and faster start up times
  • Reactive: Applicable for services with concurrent loads or longer response times
  • Stateless: Services scale better and start faster as there is no state to be passivated on shutdown or activated on start up
  • Atomic: helps to do the smallest business unit of work that can be done independently
  • Externalized Configuration: externalize the configurations in the config server, so that it can be maintained in hierarchical structure per environment
  • Consistent: Services should be written in a consistent style as per the coding standards and naming convention guidelines
  • Resilient: Service should handle exceptions arising from technical reasons (connectivity, runtime), and business reasons (invalid inputs) and not crash
  • Reporting: Usage statistics, number of times accessed, average response time, etc. via JMX API
  • Versioned: Support multiple versions for different clients, till all clients migrate to higher versions

Below is the logical view of the Microservices reference architecture.

Fig 2: Microservices Reference Architecture - Logical View

Various components of the reference architecture are described below:

Channels: Channels represents various client side or consumer applications, which will interact with Microservices.

Edge Server: API services or edge services reside on edge server or the API Gateway. Channels interact with the edge services, which decouples the microservices, and keep them channel agnostic.

API gateway is the single entry point for all clients. API Gateway is responsible for aggregating data or, in some cases, acting as a simple routing layer for appropriate services. The API gateway act as a single point of failure. For the coordination between cloud and on premise communication, an API Gateway is used.

The following diagram shows the API gateway interaction patterns

Load Balancer
It is a software based load balancer used for communication between Microservices. Configured for each service for availability, scalability and reliability.

Below diagram show load-balancing scheme.

Decentralized load balancing is the appropriate mechanism for distributing requests between available Microservice instances. Each Microservice can have its own load balancer handling only requests for that Microservice. Client is directly responsible for routing requests to an available Microservice.

API Service: API services need to be exposed either on the edge server or on API Gateway. The Service expose client specific API and can also act as coarse-grained services to orchestrate across multiple Microservices within or across bounded contexts. These services can interact with composite Microservices or directly with core Micro services.

API services can also communicate with the Enterprise Integration Infrastructure (typically ESB/MOM) to access any on premise Enterprise Applications

Composite Service: Composite services orchestrate across multiple core services. These services shall be communicating with the other services using the event-sourcing model or orchestration.

Core services: These services are the basic building blocks for Microservices architecture. These services encapsulate an entity or an aggregate (fine grained) within a given bounded context. Best practice is that, the granularity of the Microservices are always fine grained.

Circuit Breaker: Fault tolerance will ensure that when there is failure, the failed services does not adversely affect the entire system. Without proper mechanisms in place, errors, latencies will trickle up to the calling clients where they will potentially exhaust limited resources. When cascading failures occur, the overall system availability is significantly affected.

The three states of the Circuit Breaker is depicted below,

Closed State
When the service dependency is healthy and no issues detected, the Circuit Breaker is in state of closed. All invocations can pass through the service.

Open State
Circuit Breaker considers the following invocations as failed and factors them in deciding circuit open:

  • Request to the remote service time out
  • Thread pool and bounded task queue used to interact with Service dependency are at 100% capacity
  • Client library used to interact with a service dependency throws an exception

In open state, the circuit breaker reject invocations by either

  • Throwing an exception
  • Returning a fallback output

Half Open State
When the circuit breaker is in Open state, it periodically leaves through one invocation at a configurable interval. If the invocation succeeds, the circuit is closed again.

Cloud Config: It is a single source of configuration data for all other services in a Microservice-based application. Each service can have its configuration in a repository, which centralizes the configuration across all environments. It decouples the configuration from the implementation, which helps to update the configuration without affecting any of the services. Every update on the configuration files in the repository are automatically propagated to the running instances.

Service Discovery: In Microservices application, the dynamic assignment of service instances network locations happens automatically. In addition, the set of service instances changes dynamically because of auto-scaling, failures and upgrades. Consequently, the client code needs to use a service discovery mechanism.

There are two main service discovery patterns: Client-side discovery and Server-side discovery.

  • Client-Side Discovery Pattern: Client is responsible for determining the network locations of available service instances and load balancing requests across them. The client queries a service registry, which is a database of available service instances. The client then uses a load-balancing algorithm to select one of the available service instances and makes a request.
  • Server-Side Discovery Pattern: Client makes the request to a service via a load balancer. The load balancer queries the service registry and routes each request to an available service instance.
  • Service Registry: Service registry is a key part of service discovery. It is a database containing the network locations of service instances. A service registry needs to be highly available and up to date. Clients can cache network locations obtained from the service registry. However, that information eventually becomes out of date and clients become unable to discover service instances.
  • Self-Registration Pattern: Service instance is responsible for registering and deregistering itself with the service registry. In addition, a service instance sends heartbeat requests to prevent its registration from expiring.
  • Third-Party Registration Pattern: In this case, another system component known as the service registrar handles the registration. The service registrar tracks changes to the set of running instances by either polling the deployment environment or subscribing to events. It registers and deregisters the service instances.

Messaging & Events Stream: Lightweight messaging platforms like AMQP shall be used for exchange of messages between microservice within or across bounded contexts or in case of event sourcing used as part of the choreography.

Monitoring: As microservices are distributed and heterogeneous in nature, it is critical to monitor and visualize making sure that software is reliable, available, and performs as expected. Monitoring typically involves collecting metrics from all the applicable systems involved and analyzing/visualizing call graphs. These gets complex as the application complexity (again depends on number microservices and their interactions in a call graph) grows. There are many commercial and open source tools for monitoring microservices.

Distributed Tracing: Distributed tracing helps in how a request traverse through the application, especially when you may not have any insight into the implementation of the microservice you are calling. Tracing tools introduces unique IDs for logging which are consistent between Microservice calls which makes possible to find how a single request travels from one Microservice to the next.

Security: With microservices, security becomes a challenge primarily because no middleware component handles security-based functionality. Instead, each service must handle security on its own, or in some cases, the API layer be made more intelligent to handle the security aspects of the application. Security Frameworks like oAuth2 address the security concern. There are multiple ways of configuring security for microservices; by making API Gateway behave like reverse proxy or by securing each microservice using the security service (backing service like IAM) provided by the PaaS provider.

Backing Services: A PaaS provides services grouped into various categories like database, analytics, security, Data warehouse etc. used by the cloud application. These services expose lightweight protocols (like REST) and consumed by the cloud application by binding. PaaS providers provide consoles (UI based) to manage backing services.

On Premise Integration: API Gateway helps to mediate the communication between cloud and on premise applications. API Services will be used to orchestrate calls that span cloud and on premise by calling the enterprise integration infrastructure for calling any applications on premise.

Infrastructure: Infrastructure has two components; PaaS and IaaS. IaaS is the abstraction onto the hardware and provides on demand resource provisioning. Resources can be scaled out or in based on usage pattern. PaaS is cloud platform residing on top of IaaS. This layer provide the required support for the cloud applications to deploy and run. PaaS provides various runtimes and many backing services required for a cloud application.

Conclusions
Microservices is not a product, framework, or platform. It is a strategy for building large enterprise distributed systems. One of the characteristic of microservice is of loosely coupled and deployed independently of one another. Microservices architecture can offer enterprises many advantages, from independent scalability of diverse application components to faster, easier software development and maintenance. Sizing of the Microservices is very critical for the design of the better services. Open source technology solutions and organizational methods are leading the Microservices market. As a result, Microservices reduce vendor lock-in and eliminate long-term technology commitment, helps to choose the tools need to meet IT and business goals.

In addition, Microservices Reference Architecture need to be developed based on industry‑standard components like Docker containers, and a wide range of languages - Java, PHP, Python, Node.js/JavaScript and Ruby etc. Developed Microservice must expose JavaScript Object Notation (JSON) or Extensible Mark up Language (XML) over the HTTP to provide a REST API. These standards would provide guidelines about how to describe, maintain, and retire Microservices.

Finally, Microservices based System Design is an ongoing story. It is not something that has to done once, and immediately. With the right people, processes, and tools, Microservices can deliver faster development and deployment, easier maintenance, improved scalability, and freedom from long-term technology commitment.

References

  1. "Microservices". martinfowler.com. Retrieved 2017-02-06
  2. S. Newman, Building Microservices - Designing Fine-Grained Systems, O'Reilly, 2015
  3. E. Wolff, Microservices: Flexible Software Architecture, Addison-Wesley, 2016

Acknowledgements

The authors would like to thank Hari Kishan Burle, Raju Alluri of Global Enterprise Architecture

Group of Wipro Technologies for giving the required time and support in many ways in bringing

Up the article as part of Global Enterprise Architecture Practice efforts.

Authors

Dr. Gopala Krishna Behara is a Lead Enterprise Architect in the SCA Practice division of Wipro. He has a total of 21 years of IT experience. Reached at [email protected]

Tirumala Khandrika is a Senior Architect in the SCA Practice division of Wipro. He has a total of 16 years of IT experience. Reached at [email protected]

Sridhar Chalasani is an Architect in the SCA Practice division of Wipro. He has a total of 12 years of IT experience. Reached at [email protected]

Disclaimer

The views expressed in this article/presentation are that of authors and Wipro does not subscribe to the substance, veracity or truthfulness of the said opinion

More Stories By Gopala Krishna Behara

Dr. Gopala Krishna Behara is a Senior Enterprise Architect in the Enterprise Architecture & Solutions division of Wipro. He has a total of 16 years of IT experience. He can be reached at [email protected]

More Stories By Sridhar Chalasani

Sridhar Chalasani is an Architect in the Global Enterprise Architecture Practice division of Wipro. He has a total of 12 years of IT experience. He can be reached at [email protected]

More Stories By Tirumala Khandrika

Tirumala Khandrika is a Senior Architect in the Global Enterprise Architecture Practice division of Wipro. He has a total of 16 years of IT experience. He can be reached at [email protected]

@DevOpsSummit Stories
SYS-CON Events announced today that Telecom Reseller has been named “Media Sponsor” of SYS-CON's 22nd International Cloud Expo, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Telecom Reseller reports on Unified Communications, UCaaS, BPaaS for enterprise and SMBs. They report extensively on both customer premises based solutions such as IP-PBX as well as cloud based and hosted platforms.
"CA has been doing a lot of things in the area of DevOps. Now we have a complete set of tool sets in order to enable customers to go all the way from planning to development to testing down to release into the operations," explained Aruna Ravichandran, Vice President of Global Marketing and Strategy at CA Technologies, in this SYS-CON.tv interview at DevOps Summit at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
DevOps promotes continuous improvement through a culture of collaboration. But in real terms, how do you: Integrate activities across diverse teams and services? Make objective decisions with system-wide visibility? Use feedback loops to enable learning and improvement? With technology insights and real-world examples, in his general session at @DevOpsSummit, at 21st Cloud Expo, Andi Mann, Chief Technology Advocate at Splunk, explored how leading organizations use data-driven DevOps to close their feedback loops to drive continuous improvement.
While some developers care passionately about how data centers and clouds are architected, for most, it is only the end result that matters. To the majority of companies, technology exists to solve a business problem, and only delivers value when it is solving that problem. 2017 brings the mainstream adoption of containers for production workloads. In his session at 21st Cloud Expo, Ben McCormack, VP of Operations at Evernote, discussed how data centers of the future will be managed, how the public cloud best suits your organization, and what the future holds for operations and infrastructure engineers in a post-container world. Is a serverless world inevitable?
SYS-CON Events announced today that Evatronix will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Evatronix SA offers comprehensive solutions in the design and implementation of electronic systems, in CAD / CAM deployment, and also is a designer and manufacturer of advanced 3D scanners for professional applications.
SYS-CON Events announced today that CrowdReviews.com has been named “Media Sponsor” of SYS-CON's 22nd International Cloud Expo, which will take place on June 5–7, 2018, at the Javits Center in New York City, NY. CrowdReviews.com is a transparent online platform for determining which products and services are the best based on the opinion of the crowd. The crowd consists of Internet users that have experienced products and services first-hand and have an interest in letting other potential buyers learn their thoughts on their experience.
Sanjeev Sharma Joins June 5-7, 2018 @DevOpsSummit at @Cloud Expo New York Faculty. Sanjeev Sharma is an internationally known DevOps and Cloud Transformation thought leader, technology executive, and author. Sanjeev's industry experience includes tenures as CTO, Technical Sales leader, and Cloud Architect leader. As an IBM Distinguished Engineer, Sanjeev is recognized at the highest levels of IBM's core of technical leaders.
We all know that end users experience the Internet primarily with mobile devices. From an app development perspective, we know that successfully responding to the needs of mobile customers depends on rapid DevOps – failing fast, in short, until the right solution evolves in your customers' relationship to your business. Whether you’re decomposing an SOA monolith, or developing a new application cloud natively, it’s not a question of using microservices – not doing so will be a path to eventual business failure.
"Cloud4U builds software services that help people build DevOps platforms for cloud-based software and using our platform people can draw a picture of the system, network, software," explained Kihyeon Kim, CEO and Head of R&D at Cloud4U, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
A strange thing is happening along the way to the Internet of Things, namely far too many devices to work with and manage. It has become clear that we'll need much higher efficiency user experiences that can allow us to more easily and scalably work with the thousands of devices that will soon be in each of our lives. Enter the conversational interface revolution, combining bots we can literally talk with, gesture to, and even direct with our thoughts, with embedded artificial intelligence, which can process our conversational commands and orchestrate the outcomes we request across our personal and professional realm of connected devices.
Is advanced scheduling in Kubernetes achievable?Yes, however, how do you properly accommodate every real-life scenario that a Kubernetes user might encounter? How do you leverage advanced scheduling techniques to shape and describe each scenario in easy-to-use rules and configurations? In his session at @DevOpsSummit at 21st Cloud Expo, Oleg Chunikhin, CTO at Kublr, answered these questions and demonstrated techniques for implementing advanced scheduling. For example, using spot instances and cost-effective resources on AWS, coupled with the ability to deliver a minimum set of functionalities that cover the majority of needs – without configuration complexity.
As DevOps methodologies expand their reach across the enterprise, organizations face the daunting challenge of adapting related cloud strategies to ensure optimal alignment, from managing complexity to ensuring proper governance. How can culture, automation, legacy apps and even budget be reexamined to enable this ongoing shift within the modern software factory? In her Day 2 Keynote at @DevOpsSummit at 21st Cloud Expo, Aruna Ravichandran, VP, DevOps Solutions Marketing, CA Technologies, was joined by a panel of industry experts and real-world practitioners who shared their insight into an emerging set of best practices that lie at the heart of today's digital transformation.
SYS-CON Events announced today that Synametrics Technologies will exhibit at SYS-CON's 22nd International Cloud Expo®, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Synametrics Technologies is a privately held company based in Plainsboro, New Jersey that has been providing solutions for the developer community since 1997. Based on the success of its initial product offerings such as WinSQL, Xeams, SynaMan and Syncrify, Synametrics continues to create and hone innovative products that help customers get more from their computer applications, databases and infrastructure. To date, over one million users around the world have chosen Synametrics solutions to help power their accelerated business and personal computing needs.
As many know, the first generation of Cloud Management Platform (CMP) solutions were designed for managing virtual infrastructure (IaaS) and traditional applications. But that's no longer enough to satisfy evolving and complex business requirements. In his session at 21st Cloud Expo, Scott Davis, Embotics CTO, explored how next-generation CMPs ensure organizations can manage cloud-native and microservice-based application architectures, while also facilitating agile DevOps methodology. He explained how automation, orchestration and governance are fundamental to managing today's hybrid cloud environments and are critical for digital businesses to deliver services faster, with better user experience and higher quality, all while saving money.
SYS-CON Events announced today that Google Cloud has been named “Keynote Sponsor” of SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Companies come to Google Cloud to transform their businesses. Google Cloud’s comprehensive portfolio – from infrastructure to apps to devices – helps enterprises innovate faster, scale smarter, stay secure, and do more with data than ever before.
Cloud Expo | DXWorld Expo have announced the conference tracks for Cloud Expo 2018. Cloud Expo will be held June 5-7, 2018, at the Javits Center in New York City, and November 6-8, 2018, at the Santa Clara Convention Center, Santa Clara, CA. Digital Transformation (DX) is a major focus with the introduction of DX Expo within the program. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term. A total of 88% of Fortune 500 companies from a generation ago are now out of business. Only 12% still survive. Similar percentages are found throughout enterprises of all sizes.
Continuous Delivery makes it possible to exploit findings of cognitive psychology and neuroscience to increase the productivity and happiness of our teams. In his session at 22nd Cloud Expo | DXWorld Expo, Daniel Jones, CTO of EngineerBetter, will answer: How can we improve willpower and decrease technical debt? Is the present bias real? How can we turn it to our advantage? Can you increase a team’s effective IQ? How do DevOps & Product Teams increase empathy, and what impact does empathy have on productivity?
There is a huge demand for responsive, real-time mobile and web experiences, but current architectural patterns do not easily accommodate applications that respond to events in real time. Common solutions using message queues or HTTP long-polling quickly lead to resiliency, scalability and development velocity challenges. In his session at 21st Cloud Expo, Ryland Degnan, a Senior Software Engineer on the Netflix Edge Platform team, will discuss how by leveraging a reactive stream-based protocol, we have been able to solve many of these problems at the communication layer. This makes it possible to create rich application experiences and support use-cases such as mobile-to-mobile communication and large file transfers that would be difficult or cost-prohibitive with traditional networking.
You know you need the cloud, but you're hesitant to simply dump everything at Amazon since you know that not all workloads are suitable for cloud. You know that you want the kind of ease of use and scalability that you get with public cloud, but your applications are architected in a way that makes the public cloud a non-starter. You're looking at private cloud solutions based on hyperconverged infrastructure, but you're concerned with the limits inherent in those technologies. What do you do?
The 22nd International Cloud Expo | 1st DXWorld Expo has announced that its Call for Papers is open. Cloud Expo | DXWorld Expo, to be held June 5-7, 2018, at the Javits Center in New York, NY, brings together Cloud Computing, Digital Transformation, Big Data, Internet of Things, DevOps, Machine Learning and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal today!
The past few years have brought a sea change in the way applications are architected, developed, and consumed—increasing both the complexity of testing and the business impact of software failures. How can software testing professionals keep pace with modern application delivery, given the trends that impact both architectures (cloud, microservices, and APIs) and processes (DevOps, agile, and continuous delivery)? This is where continuous testing comes in. D
In his Opening Keynote at 21st Cloud Expo, John Considine, General Manager of IBM Cloud Infrastructure, led attendees through the exciting evolution of the cloud. He looked at this major disruption from the perspective of technology, business models, and what this means for enterprises of all sizes. John Considine is General Manager of Cloud Infrastructure Services at IBM. In that role he is responsible for leading IBM’s public cloud infrastructure including strategy, development, and offering management. To date, IBM has launched more than 50 cloud data centers that span the globe. He has been building advanced technology, delivering “as a service” solutions, and managing infrastructure services for the past 20 years.
Digital transformation is about embracing digital technologies into a company's culture to better connect with its customers, automate processes, create better tools, enter new markets, etc. Such a transformation requires continuous orchestration across teams and an environment based on open collaboration and daily experiments. In his session at 21st Cloud Expo, Alex Casalboni, Technical (Cloud) Evangelist at Cloud Academy, explored and discussed the most urgent unsolved challenges to achieve full cloud literacy in the enterprise world.
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the benefits of the cloud without losing performance as containers become the new paradigm.
In a recent survey, Sumo Logic surveyed 1,500 customers who employ cloud services such as Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP). According to the survey, a quarter of the respondents have already deployed Docker containers and nearly as many (23 percent) are employing the AWS Lambda serverless computing framework. It’s clear: serverless is here to stay. The adoption does come with some needed changes, within both application development and operations. That means serverless is also changing the way we leverage public clouds. Truth-be-told, many enterprise IT shops were so happy to get out of the management of physical servers within a data center that many limitations of the existing public IaaS clouds were forgiven. However, now that we’ve lived a few years with public IaaS clouds, developers and CloudOps pros are giving a huge thumbs down to the ...