Amazon EC2 Container Service vs Docker Swarm: What are the differences?
Developers describe Amazon EC2 Container Service as "Container management service that supports Docker containers". Amazon EC2 Container Service lets you launch and stop container-enabled applications with simple API calls, allows you to query the state of your cluster from a centralized service, and gives you access to many familiar Amazon EC2 features like security groups, EBS volumes and IAM roles. On the other hand, Docker Swarm is detailed as "Native clustering for Docker. Turn a pool of Docker hosts into a single, virtual host". Swarm serves the standard Docker API, so any tool which already communicates with a Docker daemon can use Swarm to transparently scale to multiple hosts: Dokku, Compose, Krane, Deis, DockerUI, Shipyard, Drone, Jenkins... and, of course, the Docker client itself.
Amazon EC2 Container Service can be classified as a tool in the "Containers as a Service" category, while Docker Swarm is grouped under "Container Tools".
"Backed by amazon" is the top reason why over 97 developers like Amazon EC2 Container Service, while over 43 developers mention "Docker friendly" as the leading cause for choosing Docker Swarm.
Docker Swarm is an open source tool with 5.61K GitHub stars and 1.11K GitHub forks. Here's a link to Docker Swarm's open source repository on GitHub.
According to the StackShare community, Amazon EC2 Container Service has a broader approval, being mentioned in 784 company stacks & 374 developers stacks; compared to Docker Swarm, which is listed in 80 company stacks and 38 developer stacks.
What is Amazon EC2 Container Service?
What is Docker Swarm?
Need advice about which tool to choose?Ask the StackShare community!
Sign up to add, upvote and see more prosMake informed product decisions
What are the cons of using Amazon EC2 Container Service?
Sign up to get full access to all the companiesMake informed product decisions
Sign up to get full access to all the tool integrationsMake informed product decisions
Visual Studio Code worked really well for us as well, it worked well with all our polyglot services and the .Net core integration had great cross-platform developer experience (to be fair, F# was a bit trickier) - actually, each of our team members used a different OS (Ubuntu, macos, windows). Our production deployment ran for a time on Docker Swarm until we've decided to adopt Kubernetes with almost seamless migration process.
After our positive experience of running .Net core workloads in containers and developing Tweek's .Net services on non-windows machines, C# had gained back some of its popularity (originally lost to Node.js), and other teams have been using it for developing microservices, k8s sidecars (like https://github.com/Soluto/airbag), cli tools, serverless functions and other projects...
We began our hosting journey, as many do, on Heroku because they make it easy to deploy your application and automate some of the routine tasks associated with deployments, etc. However, as our team grew and our product matured, our needs have outgrown Heroku. I will dive into the history and reasons for this in a future blog post.
We decided to migrate our infrastructure to Kubernetes running on Amazon EKS. Although Google Kubernetes Engine has a slightly more mature Kubernetes offering and is more user-friendly; we decided to go with EKS because we already using other AWS services (including a previous migration from Heroku Postgres to AWS RDS). We are still in the process of moving our main website workloads to EKS, however we have successfully migrate all our staging and testing PR apps to run in a staging cluster. We developed a Slack chatops application (also running in the cluster) which automates all the common tasks of spinning up and managing a production-like cluster for a pull request. This allows our engineering team to iterate quickly and safely test code in a full production environment. Helm plays a central role when deploying our staging apps into the cluster. We use CircleCI to build docker containers for each PR push, which are then published to Amazon EC2 Container Service (ECR). An
upgrade-operator process watches the ECR repository for new containers and then uses Helm to rollout updates to the staging environments. All this happens automatically and makes it really easy for developers to get code onto servers quickly. The immutable and isolated nature of our staging environments means that we can do anything we want in that environment and quickly re-create or restore the environment to start over.
The next step in our journey is to migrate our production workloads to an EKS cluster and build out the CD workflows to get our containers promoted to that cluster after our QA testing is complete in our staging environments.
We use the container service so that we can deploy our application services with Dockerfiles, so that we can test locally and deploy to AWS simply.
Additionally, the ability to scale containers and have them automatically restart in case of failure is very helpful to our operations.
We use the EC2 registry for secure private container registration. When used in combination with I AM roles we can control customer access to repos on and individual basis.
Amazon EC2 is our primary application hosting solution. Most applications are not exposed on the internet and use a virtually private cloud to interact with each other.
With a little forethought, ECS can handle a good portion of my development stack as though it were production. 12 Factor configuration makes this a breeze.