Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
It is a utility belt to handle data on AWS. It aims to fill a gap between AWS Analytics Services (Glue, Athena, EMR, Redshift) and the most popular Python data libraries (Pandas, Apache Spark). | It is a new visual data preparation tool that makes it easy for data analysts and data scientists to clean and normalize data to prepare it for analytics and machine learning. You can choose from over 250 pre-built transformations to automate data preparation tasks, all without the need to write any code. You can automate filtering anomalies, converting data to standard formats, and correcting invalid values, and other tasks. After your data is ready, you can immediately use it for analytics and machine learning projects. You only pay for what you use - no upfront commitment. |
Writes in Parquet and CSV file formats; Utility belt to handle data on AWS | Evaluate the quality of your data by profiling it to understand data patterns and detect anomalies, connect data directly from your data lake, data warehouses, and databases;
Choose from over 250 built-in transformations to visualize, clean, and normalize your data with an interactive, point-and-click visual interface;
Visually map the lineage of your data to understand the various data sources and transformation steps that the data has been through;
Automate data cleaning and normalization tasks by applying saved transformations directly to new data as it comes into your source system |
Statistics | |
Stacks 7 | Stacks 14 |
Followers 30 | Followers 13 |
Votes 0 | Votes 0 |
Integrations | |

Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more.

Besides its obvious scientific uses, NumPy can also be used as an efficient multi-dimensional container of generic data. Arbitrary data-types can be defined. This allows NumPy to seamlessly and speedily integrate with a wide variety of databases.

Integrate Python into Microsoft Excel. Use Excel as your user-facing front-end with calculations, business logic and data access powered by Python. Works with all 3rd party and open source Python packages. No need to write any VBA!

Python-based ecosystem of open-source software for mathematics, science, and engineering. It contains modules for optimization, linear algebra, integration, interpolation, special functions, FFT, signal and image processing, ODE solvers and other tasks common in science and engineering.

Dataform helps you manage all data processes in your cloud data warehouse. Publish tables, write data tests and automate complex SQL workflows in a few minutes, so you can spend more time on analytics and less time managing infrastructure.

It is the collaboration of Apache Spark and Python. it is a Python API for Spark that lets you harness the simplicity of Python and the power of Apache Spark in order to tame Big Data.

A free and open-source distribution of the Python and R programming languages for scientific computing, that aims to simplify package management and deployment. Package versions are managed by the package management system conda.

It is a versatile tool that supports a variety of workloads. It is composed of two parts: Dynamic task scheduling optimized for computation. This is similar to Airflow, Luigi, Celery, or Make, but optimized for interactive computational workloads. Big Data collections like parallel arrays, dataframes, and lists that extend common interfaces like NumPy, Pandas, or Python iterators to larger-than-memory or distributed environments. These parallel collections run on top of dynamic task schedulers.

It enable users to ingest, blend, cleanse and prepare diverse data from any source. With visual tools to eliminate coding and complexity, It puts the best quality data at the fingertips of IT and the business.

An end-to-end data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps.