Amazon ElastiCache vs Celery: What are the differences?
Developers describe Amazon ElastiCache as "Deploy, operate, and scale an in-memory cache in the cloud". ElastiCache improves the performance of web applications by allowing you to retrieve information from fast, managed, in-memory caches, instead of relying entirely on slower disk-based databases. ElastiCache supports Memcached and Redis. On the other hand, Celery is detailed as "Distributed task queue". Celery is an asynchronous task queue/job queue based on distributed message passing. It is focused on real-time operation, but supports scheduling as well.
Amazon ElastiCache belongs to "Managed Memcache" category of the tech stack, while Celery can be primarily classified under "Message Queue".
"Redis" is the top reason why over 53 developers like Amazon ElastiCache, while over 84 developers mention "Task queue" as the leading cause for choosing Celery.
Celery is an open source tool with 12.7K GitHub stars and 3.3K GitHub forks. Here's a link to Celery's open source repository on GitHub.
Instacart, SendGrid, and Sellsuki are some of the popular companies that use Amazon ElastiCache, whereas Celery is used by Sentry, Ansible, and OpenLabel. Amazon ElastiCache has a broader approval, being mentioned in 342 company stacks & 79 developers stacks; compared to Celery, which is listed in 271 company stacks and 77 developer stacks.
What is Amazon ElastiCache?
What is Celery?
Need advice about which tool to choose?Ask the StackShare community!
Sign up to add, upvote and see more prosMake informed product decisions
What are the cons of using Amazon ElastiCache?
Sign up to get full access to all the companiesMake informed product decisions
As Sentry runs throughout the day, there are about 50 different offline tasks that we execute—anything from “process this event, pretty please” to “send all of these cool people some emails.” There are some that we execute once a day and some that execute thousands per second.
Managing this variety requires a reliably high-throughput message-passing technology. We use Celery's RabbitMQ implementation, and we stumbled upon a great feature called Federation that allows us to partition our task queue across any number of RabbitMQ servers and gives us the confidence that, if any single server gets backlogged, others will pitch in and distribute some of the backlogged tasks to their consumers.
We decided to use MemCachier as our Memcached provider because we were seeing some serious PostgreSQL performance issues with query-heavy pages on the site. We use MemCachier for all Rails caching and pretty aggressively too for the logged out experience (fully cached pages for the most part). We really need to move to Amazon ElastiCache as soon as possible so we can stop paying so much. The only reason we're not moving is because there are some restrictions on the network side due to our main app being hosted on Heroku.
All of our background jobs (e.g., image resizing, file uploading, email and SMS sending) are done through Celery (using Redis as its broker). Celery's scheduling and retrying features are especially useful for error-prone tasks, such as email and SMS sending.
For orchestrating the creation of the correct number of instances, managing errors and retries, and finally managing the deallocation of resources we use RabbitMQ in conjunction with the Celery Project framework, along with a self-developed workflow engine.
We maintain a fork of Celery 3 that adds HTTPS support for Redis brokers. The Winning Model currently uses Celery 3 because Celery 4 dropped support for Windows.
We plan on migrating to Celery 4 once Azure ASE supports Linux apps
We used celery, in combination with RabbitMQ and celery-beat, to run periodic tasks, as well as some user-initiated long-running tasks on the server.
I use a micro elesticache instance as a shared session store between the Node.js clusters of dojo.zerotoherojs.com and nightly.zerotoherojs.com
Using Celery, the web service creates tasks that are executed by a background worker. Celery uses a RabbitMQ instance as a task queue.
Audit the ElastiCache configurations for best practices and standards.