Alternatives to Sidekiq logo

Alternatives to Sidekiq

Resque, Celery, RabbitMQ, delayed_job, and Kafka are the most popular alternatives and competitors to Sidekiq.
1.1K
629
+ 1
408

What is Sidekiq and what are its top alternatives?

Sidekiq is a popular open-source job scheduler for Ruby that uses Redis to manage the job queue. It offers features such as worker resilience, scheduling, web interface, and active job support. However, one limitation of Sidekiq is that it requires a Redis server to function properly.

  1. Resque: Resque is a task queuing system built on Redis, known for its simple setup and maintenance. Key features include job prioritization, multiple queues, and failure handling. Pros include high reliability and flexibility, while cons include lack of real-time monitoring.
  2. Delayed::Job: Delayed::Job is a database-backed job scheduler for Ruby applications. It offers a simple and robust way to handle background jobs. Pros include easy setup and tight integration with Rails, while cons include potential database bloat.
  3. Que: Que is a high-performance job queue for Ruby applications that uses PostgreSQL. Key features include job persistence in the database and multi-threaded worker support. Pros include low overhead and transactional job processing, while cons include limited support for other databases.
  4. Sidekiq Pro: Sidekiq Pro is the commercial version of Sidekiq with additional features such as job batch processing, priority queues, and job tagging. Pros include enhanced functionality and professional support, while cons include the need for a paid license.
  5. Sneakers: Sneakers is a fast and simple background processing framework for Ruby applications using RabbitMQ. Key features include worker concurrency control, dead-letter handling, and batch processing. Pros include high performance and RabbitMQ integration, while cons include limited Redis support.
  6. GoodJob: GoodJob is a multithreaded, Postgres-based background job queuing system for Ruby on Rails applications. It offers features such as ActiveJob compatibility, job scheduling, and real-time monitoring. Pros include simplicity and scalability, while cons include limited database support.
  7. Queuery: Queuery is a simple, Redis-based job queue library for Ruby applications. It focuses on ease of use and minimal configuration while providing essential job queuing functionality. Pros include lightweight and easy integration, while cons include limited advanced features.
  8. Shoryuken: Shoryuken is a concurrent Amazon SQS client for Ruby that enables efficient background job processing. Key features include auto-scaling, pre-fetching, and dead-letter handling. Pros include high scalability and Amazon SQS integration, while cons include limited support for other messaging systems.
  9. Sucker Punch: Sucker Punch is a simple, single-threaded background processing library for Ruby applications. It is designed for lightweight jobs that do not require complex queuing systems. Pros include simplicity and zero-configuration setup, while cons include limited scalability and reliability.
  10. Backburner: Backburner is a multi-threaded and single-threaded job processing library for Ruby applications that supports beanstalkd and other queuing systems. Key features include job priorities, delayed jobs, and job scheduling. Pros include flexibility and queuing system support, while cons include potential complexity in setup and maintenance.

Top Alternatives to Sidekiq

  • Resque
    Resque

    Background jobs can be any Ruby class or module that responds to perform. Your existing classes can easily be converted to background jobs or you can create new classes specifically to do work. Or, you can do both. ...

  • Celery
    Celery

    Celery is an asynchronous task queue/job queue based on distributed message passing. It is focused on real-time operation, but supports scheduling as well. ...

  • RabbitMQ
    RabbitMQ

    RabbitMQ gives your applications a common platform to send and receive messages, and your messages a safe place to live until received. ...

  • delayed_job
    delayed_job

    Delayed_job (or DJ) encapsulates the common pattern of asynchronously executing longer tasks in the background. It is a direct extraction from Shopify where the job table is responsible for a multitude of core tasks. ...

  • Kafka
    Kafka

    Kafka is a distributed, partitioned, replicated commit log service. It provides the functionality of a messaging system, but with a unique design. ...

  • Amazon SQS
    Amazon SQS

    Transmit any volume of data, at any level of throughput, without losing messages or requiring other services to be always available. With SQS, you can offload the administrative burden of operating and scaling a highly available messaging cluster, while paying a low price for only what you use. ...

  • Faktory
    Faktory

    Redis -> Sidekiq == Faktory -> Faktory. Faktory is a server daemon which provides a simple API to produce and consume background jobs. Jobs are a small JSON hash with a few mandatory keys. ...

  • Redis
    Redis

    Redis is an open source (BSD licensed), in-memory data structure store, used as a database, cache, and message broker. Redis provides data structures such as strings, hashes, lists, sets, sorted sets with range queries, bitmaps, hyperloglogs, geospatial indexes, and streams. ...

Sidekiq alternatives & related posts

Resque logo

Resque

118
125
9
A Redis-backed Ruby library for creating background jobs, placing them on multiple queues, and processing them later
118
125
+ 1
9
PROS OF RESQUE
  • 5
    Free
  • 3
    Scalable
  • 1
    Easy to use on heroku
CONS OF RESQUE
    Be the first to leave a con

    related Resque posts

    Celery logo

    Celery

    1.6K
    1.6K
    280
    Distributed task queue
    1.6K
    1.6K
    + 1
    280
    PROS OF CELERY
    • 99
      Task queue
    • 63
      Python integration
    • 40
      Django integration
    • 30
      Scheduled Task
    • 19
      Publish/subsribe
    • 8
      Various backend broker
    • 6
      Easy to use
    • 5
      Great community
    • 5
      Workflow
    • 4
      Free
    • 1
      Dynamic
    CONS OF CELERY
    • 4
      Sometimes loses tasks
    • 1
      Depends on broker

    related Celery posts

    James Cunningham
    Operations Engineer at Sentry · | 18 upvotes · 1.7M views
    Shared insights
    on
    CeleryCeleryRabbitMQRabbitMQ
    at

    As Sentry runs throughout the day, there are about 50 different offline tasks that we execute—anything from “process this event, pretty please” to “send all of these cool people some emails.” There are some that we execute once a day and some that execute thousands per second.

    Managing this variety requires a reliably high-throughput message-passing technology. We use Celery's RabbitMQ implementation, and we stumbled upon a great feature called Federation that allows us to partition our task queue across any number of RabbitMQ servers and gives us the confidence that, if any single server gets backlogged, others will pitch in and distribute some of the backlogged tasks to their consumers.

    #MessageQueue

    See more
    Michael Mota

    Automations are what makes a CRM powerful. With Celery and RabbitMQ we've been able to make powerful automations that truly works for our clients. Such as for example, automatic daily reports, reminders for their activities, important notifications regarding their client activities and actions on the website and more.

    We use Celery basically for everything that needs to be scheduled for the future, and using RabbitMQ as our Queue-broker is amazing since it fully integrates with Django and Celery storing on our database results of the tasks done so we can see if anything fails immediately.

    See more
    RabbitMQ logo

    RabbitMQ

    20.9K
    18.4K
    527
    Open source multiprotocol messaging broker
    20.9K
    18.4K
    + 1
    527
    PROS OF RABBITMQ
    • 234
      It's fast and it works with good metrics/monitoring
    • 79
      Ease of configuration
    • 59
      I like the admin interface
    • 50
      Easy to set-up and start with
    • 21
      Durable
    • 18
      Intuitive work through python
    • 18
      Standard protocols
    • 10
      Written primarily in Erlang
    • 8
      Simply superb
    • 6
      Completeness of messaging patterns
    • 3
      Scales to 1 million messages per second
    • 3
      Reliable
    • 2
      Distributed
    • 2
      Supports MQTT
    • 2
      Better than most traditional queue based message broker
    • 2
      Supports AMQP
    • 1
      Clusterable
    • 1
      Clear documentation with different scripting language
    • 1
      Great ui
    • 1
      Inubit Integration
    • 1
      Better routing system
    • 1
      High performance
    • 1
      Runs on Open Telecom Platform
    • 1
      Delayed messages
    • 1
      Reliability
    • 1
      Open-source
    CONS OF RABBITMQ
    • 9
      Too complicated cluster/HA config and management
    • 6
      Needs Erlang runtime. Need ops good with Erlang runtime
    • 5
      Configuration must be done first, not by your code
    • 4
      Slow

    related RabbitMQ posts

    James Cunningham
    Operations Engineer at Sentry · | 18 upvotes · 1.7M views
    Shared insights
    on
    CeleryCeleryRabbitMQRabbitMQ
    at

    As Sentry runs throughout the day, there are about 50 different offline tasks that we execute—anything from “process this event, pretty please” to “send all of these cool people some emails.” There are some that we execute once a day and some that execute thousands per second.

    Managing this variety requires a reliably high-throughput message-passing technology. We use Celery's RabbitMQ implementation, and we stumbled upon a great feature called Federation that allows us to partition our task queue across any number of RabbitMQ servers and gives us the confidence that, if any single server gets backlogged, others will pitch in and distribute some of the backlogged tasks to their consumers.

    #MessageQueue

    See more

    Around the time of their Series A, Pinterest’s stack included Python and Django, with Tornado and Node.js as web servers. Memcached / Membase and Redis handled caching, with RabbitMQ handling queueing. Nginx, HAproxy and Varnish managed static-delivery and load-balancing, with persistent data storage handled by MySQL.

    See more
    delayed_job logo

    delayed_job

    51
    64
    6
    Database backed asynchronous priority queue -- Extracted from Shopify
    51
    64
    + 1
    6
    PROS OF DELAYED_JOB
    • 3
      Easy to get started
    • 2
      Reliable
    • 1
      Doesn't require Redis
    CONS OF DELAYED_JOB
      Be the first to leave a con

      related delayed_job posts

      Jerome Dalbert
      Principal Backend Software Engineer at StackShare · | 4 upvotes · 86.6K views

      delayed_job is a great Rails background job library for new projects, as it only uses what you already have: a relational database. We happily used it during the company’s first two years.

      But it started to falter as our web and database transactions significantly grew. Our app interacted with users via SMS texts sent inside background jobs. Because the delayed_job daemon ran every couple seconds, this meant that users often waited several long seconds before getting text replies, which was not acceptable. Moreover, job processing was done inside AWS Elastic Beanstalk web instances, which were already under stress and not meant to handle jobs.

      We needed a fast background job system that could process jobs in near real-time and integrate well with AWS. Sidekiq is a fast and popular Ruby background job library, but it does not leverage the Elastic Beanstalk worker architecture, and you have to maintain a Redis instance.

      We ended up choosing active-elastic-job, which seamlessly integrates with worker instances and Amazon SQS. SQS is a fast queue and you don’t need to worry about infrastructure or scaling, as AWS handles it for you.

      We noticed significant performance gains immediately after making the switch.

      #BackgroundProcessing

      See more
      Jerome Dalbert
      Principal Backend Software Engineer at StackShare · | 3 upvotes · 62.3K views

      We use Sidekiq to process millions of Ruby background jobs a day under normal loads. We sometimes process more than that when running one-off backfill tasks.

      With so many jobs, it wouldn't really make sense to use delayed_job, as it would put our main database under unnecessary load, which would make it a bottleneck with most DB queries serving jobs and not end users. I suppose you could create a separate DB just for jobs, but that can be a hassle. Sidekiq uses a separate Redis instance so you don't have this problem. And it is very performant!

      I also like that its free version comes "batteries included" with:

      • A web monitoring UI that provides some nice stats.
      • An API that can come in handy for one-off tasks, like changing the queue of certain already enqueued jobs.

      Sidekiq is a pleasure to use. All our engineers love it!

      See more
      Kafka logo

      Kafka

      23K
      21.6K
      607
      Distributed, fault tolerant, high throughput pub-sub messaging system
      23K
      21.6K
      + 1
      607
      PROS OF KAFKA
      • 126
        High-throughput
      • 119
        Distributed
      • 92
        Scalable
      • 86
        High-Performance
      • 66
        Durable
      • 38
        Publish-Subscribe
      • 19
        Simple-to-use
      • 18
        Open source
      • 12
        Written in Scala and java. Runs on JVM
      • 9
        Message broker + Streaming system
      • 4
        KSQL
      • 4
        Avro schema integration
      • 4
        Robust
      • 3
        Suport Multiple clients
      • 2
        Extremely good parallelism constructs
      • 2
        Partioned, replayable log
      • 1
        Simple publisher / multi-subscriber model
      • 1
        Fun
      • 1
        Flexible
      CONS OF KAFKA
      • 32
        Non-Java clients are second-class citizens
      • 29
        Needs Zookeeper
      • 9
        Operational difficulties
      • 5
        Terrible Packaging

      related Kafka posts

      Eric Colson
      Chief Algorithms Officer at Stitch Fix · | 21 upvotes · 6.1M views

      The algorithms and data infrastructure at Stitch Fix is housed in #AWS. Data acquisition is split between events flowing through Kafka, and periodic snapshots of PostgreSQL DBs. We store data in an Amazon S3 based data warehouse. Apache Spark on Yarn is our tool of choice for data movement and #ETL. Because our storage layer (s3) is decoupled from our processing layer, we are able to scale our compute environment very elastically. We have several semi-permanent, autoscaling Yarn clusters running to serve our data processing needs. While the bulk of our compute infrastructure is dedicated to algorithmic processing, we also implemented Presto for adhoc queries and dashboards.

      Beyond data movement and ETL, most #ML centric jobs (e.g. model training and execution) run in a similarly elastic environment as containers running Python and R code on Amazon EC2 Container Service clusters. The execution of batch jobs on top of ECS is managed by Flotilla, a service we built in house and open sourced (see https://github.com/stitchfix/flotilla-os).

      At Stitch Fix, algorithmic integrations are pervasive across the business. We have dozens of data products actively integrated systems. That requires serving layer that is robust, agile, flexible, and allows for self-service. Models produced on Flotilla are packaged for deployment in production using Khan, another framework we've developed internally. Khan provides our data scientists the ability to quickly productionize those models they've developed with open source frameworks in Python 3 (e.g. PyTorch, sklearn), by automatically packaging them as Docker containers and deploying to Amazon ECS. This provides our data scientist a one-click method of getting from their algorithms to production. We then integrate those deployments into a service mesh, which allows us to A/B test various implementations in our product.

      For more info:

      #DataScience #DataStack #Data

      See more
      John Kodumal

      As we've evolved or added additional infrastructure to our stack, we've biased towards managed services. Most new backing stores are Amazon RDS instances now. We do use self-managed PostgreSQL with TimescaleDB for time-series data—this is made HA with the use of Patroni and Consul.

      We also use managed Amazon ElastiCache instances instead of spinning up Amazon EC2 instances to run Redis workloads, as well as shifting to Amazon Kinesis instead of Kafka.

      See more
      Amazon SQS logo

      Amazon SQS

      2.2K
      2K
      171
      Fully managed message queuing service
      2.2K
      2K
      + 1
      171
      PROS OF AMAZON SQS
      • 62
        Easy to use, reliable
      • 40
        Low cost
      • 28
        Simple
      • 14
        Doesn't need to maintain it
      • 8
        It is Serverless
      • 4
        Has a max message size (currently 256K)
      • 3
        Triggers Lambda
      • 3
        Easy to configure with Terraform
      • 3
        Delayed delivery upto 15 mins only
      • 3
        Delayed delivery upto 12 hours
      • 1
        JMS compliant
      • 1
        Support for retry and dead letter queue
      • 1
        D
      CONS OF AMAZON SQS
      • 2
        Has a max message size (currently 256K)
      • 2
        Proprietary
      • 2
        Difficult to configure
      • 1
        Has a maximum 15 minutes of delayed messages only

      related Amazon SQS posts

      Praveen Mooli
      Engineering Manager at Taylor and Francis · | 18 upvotes · 3.8M views

      We are in the process of building a modern content platform to deliver our content through various channels. We decided to go with Microservices architecture as we wanted scale. Microservice architecture style is an approach to developing an application as a suite of small independently deployable services built around specific business capabilities. You can gain modularity, extensive parallelism and cost-effective scaling by deploying services across many distributed servers. Microservices modularity facilitates independent updates/deployments, and helps to avoid single point of failure, which can help prevent large-scale outages. We also decided to use Event Driven Architecture pattern which is a popular distributed asynchronous architecture pattern used to produce highly scalable applications. The event-driven architecture is made up of highly decoupled, single-purpose event processing components that asynchronously receive and process events.

      To build our #Backend capabilities we decided to use the following: 1. #Microservices - Java with Spring Boot , Node.js with ExpressJS and Python with Flask 2. #Eventsourcingframework - Amazon Kinesis , Amazon Kinesis Firehose , Amazon SNS , Amazon SQS, AWS Lambda 3. #Data - Amazon RDS , Amazon DynamoDB , Amazon S3 , MongoDB Atlas

      To build #Webapps we decided to use Angular 2 with RxJS

      #Devops - GitHub , Travis CI , Terraform , Docker , Serverless

      See more
      Tim Specht
      ‎Co-Founder and CTO at Dubsmash · | 14 upvotes · 938.9K views

      In order to accurately measure & track user behaviour on our platform we moved over quickly from the initial solution using Google Analytics to a custom-built one due to resource & pricing concerns we had.

      While this does sound complicated, it’s as easy as clients sending JSON blobs of events to Amazon Kinesis from where we use AWS Lambda & Amazon SQS to batch and process incoming events and then ingest them into Google BigQuery. Once events are stored in BigQuery (which usually only takes a second from the time the client sends the data until it’s available), we can use almost-standard-SQL to simply query for data while Google makes sure that, even with terabytes of data being scanned, query times stay in the range of seconds rather than hours. Before ingesting their data into the pipeline, our mobile clients are aggregating events internally and, once a certain threshold is reached or the app is going to the background, sending the events as a JSON blob into the stream.

      In the past we had workers running that continuously read from the stream and would validate and post-process the data and then enqueue them for other workers to write them to BigQuery. We went ahead and implemented the Lambda-based approach in such a way that Lambda functions would automatically be triggered for incoming records, pre-aggregate events, and write them back to SQS, from which we then read them, and persist the events to BigQuery. While this approach had a couple of bumps on the road, like re-triggering functions asynchronously to keep up with the stream and proper batch sizes, we finally managed to get it running in a reliable way and are very happy with this solution today.

      #ServerlessTaskProcessing #GeneralAnalytics #RealTimeDataProcessing #BigDataAsAService

      See more
      Faktory logo

      Faktory

      6
      27
      3
      Background jobs for any language, by the makers of Sidekiq
      6
      27
      + 1
      3
      PROS OF FAKTORY
      • 2
        Worker language agnostic
      • 1
        Simple service API
      CONS OF FAKTORY
        Be the first to leave a con

        related Faktory posts

        Redis logo

        Redis

        58.2K
        44.8K
        3.9K
        Open source (BSD licensed), in-memory data structure store
        58.2K
        44.8K
        + 1
        3.9K
        PROS OF REDIS
        • 886
          Performance
        • 542
          Super fast
        • 513
          Ease of use
        • 444
          In-memory cache
        • 324
          Advanced key-value cache
        • 194
          Open source
        • 182
          Easy to deploy
        • 164
          Stable
        • 155
          Free
        • 121
          Fast
        • 42
          High-Performance
        • 40
          High Availability
        • 35
          Data Structures
        • 32
          Very Scalable
        • 24
          Replication
        • 22
          Great community
        • 22
          Pub/Sub
        • 19
          "NoSQL" key-value data store
        • 16
          Hashes
        • 13
          Sets
        • 11
          Sorted Sets
        • 10
          NoSQL
        • 10
          Lists
        • 9
          Async replication
        • 9
          BSD licensed
        • 8
          Bitmaps
        • 8
          Integrates super easy with Sidekiq for Rails background
        • 7
          Keys with a limited time-to-live
        • 7
          Open Source
        • 6
          Lua scripting
        • 6
          Strings
        • 5
          Awesomeness for Free
        • 5
          Hyperloglogs
        • 4
          Transactions
        • 4
          Outstanding performance
        • 4
          Runs server side LUA
        • 4
          LRU eviction of keys
        • 4
          Feature Rich
        • 4
          Written in ANSI C
        • 4
          Networked
        • 3
          Data structure server
        • 3
          Performance & ease of use
        • 2
          Dont save data if no subscribers are found
        • 2
          Automatic failover
        • 2
          Easy to use
        • 2
          Temporarily kept on disk
        • 2
          Scalable
        • 2
          Existing Laravel Integration
        • 2
          Channels concept
        • 2
          Object [key/value] size each 500 MB
        • 2
          Simple
        CONS OF REDIS
        • 15
          Cannot query objects directly
        • 3
          No secondary indexes for non-numeric data types
        • 1
          No WAL

        related Redis posts

        Robert Zuber

        We use MongoDB as our primary #datastore. Mongo's approach to replica sets enables some fantastic patterns for operations like maintenance, backups, and #ETL.

        As we pull #microservices from our #monolith, we are taking the opportunity to build them with their own datastores using PostgreSQL. We also use Redis to cache data we’d never store permanently, and to rate-limit our requests to partners’ APIs (like GitHub).

        When we’re dealing with large blobs of immutable data (logs, artifacts, and test results), we store them in Amazon S3. We handle any side-effects of S3’s eventual consistency model within our own code. This ensures that we deal with user requests correctly while writes are in process.

        See more

        I'm working as one of the engineering leads in RunaHR. As our platform is a Saas, we thought It'd be good to have an API (We chose Ruby and Rails for this) and a SPA (built with React and Redux ) connected. We started the SPA with Create React App since It's pretty easy to start.

        We use Jest as the testing framework and react-testing-library to test React components. In Rails we make tests using RSpec.

        Our main database is PostgreSQL, but we also use MongoDB to store some type of data. We started to use Redis  for cache and other time sensitive operations.

        We have a couple of extra projects: One is an Employee app built with React Native and the other is an internal back office dashboard built with Next.js for the client and Python in the backend side.

        Since we have different frontend apps we have found useful to have Bit to document visual components and utils in JavaScript.

        See more