Amazon EKS vs AWS Fargate: What are the differences?
Amazon EKS: Highly available and scalable Kubernetes service. Amazon Elastic Container Service for Kubernetes (Amazon EKS) is a managed service that makes it easy for you to run Kubernetes on AWS without needing to install and operate your own Kubernetes clusters; AWS Fargate: Run Containers Without Managing Infrastructure. AWS Fargate is a technology for Amazon ECS and EKS* that allows you to run containers without having to manage servers or clusters. With AWS Fargate, you no longer have to provision, configure, and scale clusters of virtual machines to run containers.
Amazon EKS and AWS Fargate can be primarily classified as "Containers as a Service" tools.
Some of the features offered by Amazon EKS are:
- Managed Kubernetes Control Plane - Amazon EKS provides a scalable and highly-available control plane that runs across multiple AWS availability zones.
- Security and Networking - Amazon EKS makes it easy to provide security for your Kubernetes clusters, with advanced features and integrations to AWS services and technology partner solutions.
- Logging - Amazon EKS is integrated with Amazon CloudWatch Logs and AWS CloudTrail to provide visibility and audit history tracking of your cluster and user activity.
On the other hand, AWS Fargate provides the following key features:
- No clusters to manage
- seamless scaling
- integrated with Amazon ECS and EKS
Third Iron, MAK IT, and Even Financial are some of the popular companies that use AWS Fargate, whereas Amazon EKS is used by Moogsoft, Condé Nast, and Cuemby. AWS Fargate has a broader approval, being mentioned in 36 company stacks & 12 developers stacks; compared to Amazon EKS, which is listed in 25 company stacks and 7 developer stacks.
What is Amazon EKS?
What is AWS Fargate?
Need advice about which tool to choose?Ask the StackShare community!
Why do developers choose Amazon EKS?
Why do developers choose AWS Fargate?
What are the cons of using Amazon EKS?
What are the cons of using AWS Fargate?
Sign up to get full access to all the companiesMake informed product decisions
Sign up to get full access to all the tool integrationsMake informed product decisions
We began our hosting journey, as many do, on Heroku because they make it easy to deploy your application and automate some of the routine tasks associated with deployments, etc. However, as our team grew and our product matured, our needs have outgrown Heroku. I will dive into the history and reasons for this in a future blog post.
We decided to migrate our infrastructure to Kubernetes running on Amazon EKS. Although Google Kubernetes Engine has a slightly more mature Kubernetes offering and is more user-friendly; we decided to go with EKS because we already using other AWS services (including a previous migration from Heroku Postgres to AWS RDS). We are still in the process of moving our main website workloads to EKS, however we have successfully migrate all our staging and testing PR apps to run in a staging cluster. We developed a Slack chatops application (also running in the cluster) which automates all the common tasks of spinning up and managing a production-like cluster for a pull request. This allows our engineering team to iterate quickly and safely test code in a full production environment. Helm plays a central role when deploying our staging apps into the cluster. We use CircleCI to build docker containers for each PR push, which are then published to Amazon EC2 Container Service (ECR). An
upgrade-operator process watches the ECR repository for new containers and then uses Helm to rollout updates to the staging environments. All this happens automatically and makes it really easy for developers to get code onto servers quickly. The immutable and isolated nature of our staging environments means that we can do anything we want in that environment and quickly re-create or restore the environment to start over.
The next step in our journey is to migrate our production workloads to an EKS cluster and build out the CD workflows to get our containers promoted to that cluster after our QA testing is complete in our staging environments.