Druid vs HBase: What are the differences?
What is Druid? Fast column-oriented distributed data store. Druid is a distributed, column-oriented, real-time analytics data store that is commonly used to power exploratory dashboards in multi-tenant environments. Druid excels as a data warehousing solution for fast aggregate queries on petabyte sized data sets. Druid supports a variety of flexible filters, exact calculations, approximate algorithms, and other useful calculations.
What is HBase? The Hadoop database, a distributed, scalable, big data store. Apache HBase is an open-source, distributed, versioned, column-oriented store modeled after Google' Bigtable: A Distributed Storage System for Structured Data by Chang et al. Just as Bigtable leverages the distributed data storage provided by the Google File System, HBase provides Bigtable-like capabilities on top of Apache Hadoop.
Druid belongs to "Big Data Tools" category of the tech stack, while HBase can be primarily classified under "Databases".
"Real Time Aggregations" is the primary reason why developers consider Druid over the competitors, whereas "Performance" was stated as the key factor in picking HBase.
Druid and HBase are both open source tools. Druid with 8.31K GitHub stars and 2.08K forks on GitHub appears to be more popular than HBase with 2.91K GitHub stars and 2.01K GitHub forks.
Pinterest, HubSpot, and hike are some of the popular companies that use HBase, whereas Druid is used by Airbnb, Instacart, and Dial Once. HBase has a broader approval, being mentioned in 54 company stacks & 18 developers stacks; compared to Druid, which is listed in 24 company stacks and 12 developer stacks.
What is Druid?
What is HBase?
Need advice about which tool to choose?Ask the StackShare community!
Sign up to add, upvote and see more prosMake informed product decisions
What are the cons of using Druid?
What are the cons of using HBase?
Sign up to get full access to all the companiesMake informed product decisions
Sign up to get full access to all the tool integrationsMake informed product decisions
The final output is inserted into HBase to serve the experiment dashboard. We also load the output data to Redshift for ad-hoc analysis. For real-time experiment data processing, we use Storm to tail Kafka and process data in real-time and insert metrics into MySQL, so we could identify group allocation problems and send out real-time alerts and metrics.