The document discusses migrating to cloud native solutions. It defines cloud native as an approach that exploits the advantages of cloud computing using containers, microservices, and other modern technologies. This allows applications to be scalable, resilient, and manageable. The document outlines the benefits of cloud native and provides a "trail map" to transitioning applications. It also discusses common challenges like technical debt and failing to meet CI/CD expectations, and provides recommendations to address them such as automating processes and simplifying architectures.
We hear a lot about using service mesh with Kubernetes and public clouds, but what about outside the clouds? In this talk, you’ll learn creative ways to apply a service mesh across different platforms and environments to automate canary deployments, facilitate cloud migrations, and more. By combining HashiCorp Consul’s service mesh and Terraform’s infrastructure as code, you can build a more seamless operational experience across multiple environments.
Join Datadog for a webinar on monitoring Kubernetes with a focus on Amazon EKS. You'll learn how to get the most out of Datadog's intuitive platform and EKS's unique capabilities, including: How to monitor metrics, logs and traces from your EKS environment How to test the usability of your environment with features such as adaptive Browser Tests and globally available Real User Monitoring How to find and fix user-facing issues with synthetic monitoring features like adaptive Browser Tests and globally available Real User Monitoring
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs, and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud native apps. But what to do if you’ve no shiny new cloud native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can! We’re facing the challenge of migrating hundreds of JEE legacy applications of a German blue chip company onto a Kubernetes cluster within one year. The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way.
The document discusses the emerging "cloud-native" ecosystem centered around containers. It identifies key characteristics like containers as modular compute units and microservices architectures. Popular early solutions are mentioned like Docker, CoreOS, Kubernetes, and Mesosphere, but the ecosystem remains immature with issues around persistence, security, and lack of best practices. Standards are emerging that may drive further innovation, and containers still lack a "killer app" business case like virtualization had with consolidation. The document provides a taxonomy of the technology stack and lists many active companies and projects in different layers.
An important use-case for Vault is to provide short lived and least privileged Cloud credentials. In this webinar we will review specifically how Vault's Azure Secrets Engine can provide dynamic Azure credentials. We will cover details on how to configure the Azure Secrets Engine in Vault and use it in an application. If you are using Azure now or in the near future, join us for some patterns on maintaining a high security posture with Vault's dynamic credentials model!
Operations practices have historically lagged behind development. Agile and Extreme Programming have become common practice for development teams. In the last decade, the DevOps and SRE movements have brought these concepts to operations, borrowing heavily from Lean principles such as Kanban and Value Stream Mapping. So, how does all of this play out if we’re using Kubernetes? In this class, Paul Czarkowski, Principal Technologist at Pivotal, will explain how Kubernetes enables a new cloud-native way of operating software. Attend to learn: ● what cloud-native operations are; ● how to build a cloud-native CI/CD stack; and ● how to deploy and upgrade an application from source to production on Kubernetes. Presenter: Paul Czarkowski, Principal Technologist, Pivotal Software
Everyone is talking about Containers, but what is this really about what are the benefits of Containers for your customers? You probably think you know, but there is more! And did you know you can run and manage Containers in the Microsoft Cloud? This session will go in to the benefits of Containers for your customers and what Microsoft is offering to facilitate in all your needs. We will touch on technologies like Kubernetes, Docker and we will elaborate on the strong partnerships Microsoft has built with true Open Source companies like Red Hat.
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs, and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud native apps. But what to do if you’ve no shiny new cloud native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can! We’re facing the challenge of migrating hundreds of JEE legacy applications of a major German insurance company onto a Kubernetes cluster within one year. We're now close to the finish line and it worked pretty well so far. The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way. We'll provide our answers to life, the universe and a cloud native journey like: - What technical constraints of Kubernetes can be obstacles for applications and how to tackle these? - How to architect a landscape of hundreds of containerized applications with their surrounding infrastructure like DBs MQs and IAM and heavy requirements on security? - How to industrialize and govern the migration process? - How to leverage the possibilities of a cloud native platform like Kubernetes without challenging the tight timeline?
Broadridge migrated their DevOps tools from an on-premise installation to AWS using Kubernetes. They prepared by gaining Kubernetes expertise, setting expectations that it would take time, and building a minimal viable product. Their process involved documenting and testing the architecture, using infrastructure as code, and leveraging third party expertise. Results included reusable Kubernetes and Helm configurations. Going forward, they will migrate more customers to the new platform and automate testing.
As a cloud native application grows in size—more microservices, more dependencies, more teams—there’s a corresponding increase in… Complexity: Over time, the application becomes a lot harder for a single developer to reason about and contribute to. Staying on top of READMEs and managing cross-team communication is practically a full-time job. Scaling challenges: The reality of building, deploying, and testing a 100+ service distributed application means developers are going to spend a lot of time sitting around waiting. But it doesn’t have to end up this way, and there are concrete steps that DevOps engineers can take to keep their developers moving quickly even as an application grows. In this webinar, we’ll show you how to use open source products to: Make it easy for your developers to code and run on-demand tests against a production-like environment—without having to constantly deal with the complexity that comes with a large application Codify the relationship between all your services and tests, making your system self-documented and easy to understand Keep your integration tests running fast so that devs can more easily write and debug their tests and get the quick feedback loops they need Facilitate remote, in-cluster development and give every developer their own isolated namespace—and never again ask a developer to deploy the application on their laptop
This document discusses Syncier's approach to migrating 152 legacy applications from a traditional platform to a cloud native platform within 17 months. Key aspects included: 1) All applications were containerized using a 12-factor app approach and migrated to production on the cloud platform. 2) The applications were modernized and security hardened during the migration process. 3) The migration provided benefits like higher availability, more agile development teams, and a strong business case. The document then discusses perspectives from an architect, project manager, and developer on the migration process. It emphasizes an emergent design approach, industrialization of the migration through training and support, and transparency throughout the project.
Introduction to Istio service mesh (e.g., history, context, features and roadmap) Presented at the Tel Aviv cloud meetup