CouchDB vs Hadoop

Need advice about which tool to choose?Ask the StackShare community!

CouchDB

429
500
+ 1
139
Hadoop

2K
2K
+ 1
55
Add tool

CouchDB vs Hadoop: What are the differences?

Developers describe CouchDB as "HTTP + JSON document database with Map Reduce views and peer-based replication". Apache CouchDB is a database that uses JSON for documents, JavaScript for MapReduce indexes, and regular HTTP for its API. CouchDB is a database that completely embraces the web. Store your data with JSON documents. Access your documents and query your indexes with your web browser, via HTTP. Index, combine, and transform your documents with JavaScript. On the other hand, Hadoop is detailed as "Open-source software for reliable, scalable, distributed computing". The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.

CouchDB and Hadoop can be categorized as "Databases" tools.

"JSON" is the primary reason why developers consider CouchDB over the competitors, whereas "Great ecosystem" was stated as the key factor in picking Hadoop.

CouchDB and Hadoop are both open source tools. Hadoop with 9.18K GitHub stars and 5.74K forks on GitHub appears to be more popular than CouchDB with 4.22K GitHub stars and 833 GitHub forks.

Slack, Shopify, and SendGrid are some of the popular companies that use Hadoop, whereas CouchDB is used by BrightMachine, Third Iron, and SocialDecode. Hadoop has a broader approval, being mentioned in 237 company stacks & 116 developers stacks; compared to CouchDB, which is listed in 60 company stacks and 30 developer stacks.

Advice on CouchDB and Hadoop
Needs advice
on
SnowflakeSnowflakeMarkLogicMarkLogic
and
HadoopHadoop

For a property and casualty insurance company, we currently use MarkLogic and Hadoop for our raw data lake. Trying to figure out how snowflake fits in the picture. Does anybody have some good suggestions/best practices for when to use and what data to store in Mark logic versus Snowflake versus a hadoop or all three of these platforms redundant with one another?

See more
Needs advice
on
SnowflakeSnowflakeMarkLogicMarkLogic
and
HadoopHadoop

for property and casualty insurance company we current Use marklogic and Hadoop for our raw data lake. Trying to figure out how snowflake fits in the picture. Does anybody have some good suggestions/best practices for when to use and what data to store in Mark logic versus snowflake versus a hadoop or all three of these platforms redundant with one another?

See more
Replies (1)
Ivo Dinis Rodrigues
none of you bussines at Marklogic · | 1 upvotes · 4.1K views
Recommends

As i see it, you can use Snowflake as your data warehouse and marklogic as a data lake. You can add all your raw data to ML and curate it to a company data model to then supply this to Snowflake. You could try to implement the dw functionality on marklogic but it will just cost you alot of time. If you are using Aws version of Snowflake you can use ML spark connector to access the data. As an extra you can use the ML also as an Operational report system if you join it with a Reporting tool lie PowerBi. With extra apis you can also provide data to other systems with ML as source.

See more
Needs advice
on
KafkaKafkaInfluxDBInfluxDB
and
HadoopHadoop

I have a lot of data that's currently sitting in a MariaDB database, a lot of tables that weigh 200gb with indexes. Most of the large tables have a date column which is always filtered, but there are usually 4-6 additional columns that are filtered and used for statistics. I'm trying to figure out the best tool for storing and analyzing large amounts of data. Preferably self-hosted or a cheap solution. The current problem I'm running into is speed. Even with pretty good indexes, if I'm trying to load a large dataset, it's pretty slow.

See more
Replies (1)
Recommends
DruidDruid

Druid Could be an amazing solution for your use case, My understanding, and the assumption is you are looking to export your data from MariaDB for Analytical workload. It can be used for time series database as well as a data warehouse and can be scaled horizontally once your data increases. It's pretty easy to set up on any environment (Cloud, Kubernetes, or Self-hosted nix system). Some important features which make it a perfect solution for your use case. 1. It can do streaming ingestion (Kafka, Kinesis) as well as batch ingestion (Files from Local & Cloud Storage or Databases like MySQL, Postgres). In your case MariaDB (which has the same drivers to MySQL) 2. Columnar Database, So you can query just the fields which are required, and that runs your query faster automatically. 3. Druid intelligently partitions data based on time and time-based queries are significantly faster than traditional databases. 4. Scale up or down by just adding or removing servers, and Druid automatically rebalances. Fault-tolerant architecture routes around server failures 5. Gives ana amazing centralized UI to manage data sources, query, tasks.

See more
Decisions about CouchDB and Hadoop
James Bender
Lead Application Architect at TekPartners · | 4 upvotes · 488 views

Our application data all goes in SQL. We will use something like Cosmos or Couch DB if one or both of these conditions are true: *We need to ingest a large amount of bulk data from a third party, and integrating it straight into an RDBMS with referential integrity checks would create a performance hit *We need to ingest a large amount of data that does not have a clearly defined, or consistent schema. In either case, we will have a process that migrates the data from Cosmos/Couch to SQL in a way that doesn't create a noticeable performance hit and ensures that we are not introducing bad data to the system. Because of this, there is a third condition that must be met: the data that is coming in must be something that the users will not need immediately, i.e. stock ticker information, real-time telemetry from other systems for performance/safety monitoring, etc.

See more
Gabriel Pa

We implemented our first large scale EPR application from naologic.com using CouchDB .

Very fast, replication works great, doesn't consume much RAM, queries are blazing fast but we found a problem: the queries were very hard to write, it took a long time to figure out the API, we had to go and write our own @nodejs library to make it work properly.

It lost most of its support. Since then, we migrated to Couchbase and the learning curve was steep but all worth it. Memcached indexing out of the box, full text search works great.

See more
Get Advice from developers at your company using Private StackShare. Sign up for Private StackShare.
Learn More
Pros of CouchDB
Pros of Hadoop
  • 43
    JSON
  • 30
    Open source
  • 18
    Highly available
  • 12
    Partition tolerant
  • 11
    Eventual consistency
  • 7
    Sync
  • 5
    REST API
  • 4
    Attachments mechanism to docs
  • 4
    Multi master replication
  • 3
    Changes feed
  • 1
    REST interface
  • 1
    js- and erlang-views
  • 38
    Great ecosystem
  • 11
    One stack to rule them all
  • 4
    Great load balancer
  • 1
    Amazon aws
  • 1
    Java syntax

Sign up to add or upvote prosMake informed product decisions

What is CouchDB?

Apache CouchDB is a database that uses JSON for documents, JavaScript for MapReduce indexes, and regular HTTP for its API. CouchDB is a database that completely embraces the web. Store your data with JSON documents. Access your documents and query your indexes with your web browser, via HTTP. Index, combine, and transform your documents with JavaScript.

What is Hadoop?

The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.

Need advice about which tool to choose?Ask the StackShare community!

What companies use CouchDB?
What companies use Hadoop?
See which teams inside your own company are using CouchDB or Hadoop.
Sign up for Private StackShareLearn More

Sign up to get full access to all the companiesMake informed product decisions

What tools integrate with CouchDB?
What tools integrate with Hadoop?

Sign up to get full access to all the tool integrationsMake informed product decisions

Blog Posts

MySQLKafkaApache Spark+6
2
1590
Aug 28 2019 at 3:10AM

Segment

PythonJavaAmazon S3+16
5
2150
What are some alternatives to CouchDB and Hadoop?
MongoDB
MongoDB stores data in JSON-like documents that can vary in structure, offering a dynamic, flexible schema. MongoDB was also designed for high availability and scalability, with built-in replication and auto-sharding.
Couchbase
Developed as an alternative to traditionally inflexible SQL databases, the Couchbase NoSQL database is built on an open source foundation and architected to help developers solve real-world problems and meet high scalability demands.
Cloudant
Cloudant’s distributed database as a service (DBaaS) allows developers of fast-growing web and mobile apps to focus on building and improving their products, instead of worrying about scaling and managing databases on their own.
MariaDB
Started by core members of the original MySQL team, MariaDB actively works with outside developers to deliver the most featureful, stable, and sanely licensed open SQL server in the industry. MariaDB is designed as a drop-in replacement of MySQL(R) with more features, new storage engines, fewer bugs, and better performance.
RethinkDB
RethinkDB is built to store JSON documents, and scale to multiple machines with very little effort. It has a pleasant query language that supports really useful queries like table joins and group by, and is easy to setup and learn.
See all alternatives