Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more. | A free and open-source distribution of the Python and R programming languages for scientific computing, that aims to simplify package management and deployment. Package versions are managed by the package management system conda. |
Easy handling of missing data (represented as NaN) in floating point as well as non-floating point data;Size mutability: columns can be inserted and deleted from DataFrame and higher dimensional objects;Automatic and explicit data alignment: objects can be explicitly aligned to a set of labels, or the user can simply ignore the labels and let Series, DataFrame, etc. automatically align the data for you in computations;Powerful, flexible group by functionality to perform split-apply-combine operations on data sets, for both aggregating and transforming data;Make it easy to convert ragged, differently-indexed data in other Python and NumPy data structures into DataFrame objects;Intelligent label-based slicing, fancy indexing, and subsetting of large data sets;Intuitive merging and joining data sets;Flexible reshaping and pivoting of data sets;Hierarchical labeling of axes (possible to have multiple labels per tick);Robust IO tools for loading data from flat files (CSV and delimited), Excel files, databases, and saving/loading data from the ultrafast HDF5 format;Time series-specific functionality: date range generation and frequency conversion, moving window statistics, moving window linear regressions, date shifting and lagging, etc. | Stay safe and secure;
Deliver on your data strategy;
Get to market faster;
Maximize flexibility and control |
Statistics | |
Stacks 1.8K | Stacks 437 |
Followers 1.3K | Followers 490 |
Votes 23 | Votes 0 |
Pros & Cons | |
Pros
| No community feedback yet |
Integrations | |

Besides its obvious scientific uses, NumPy can also be used as an efficient multi-dimensional container of generic data. Arbitrary data-types can be defined. This allows NumPy to seamlessly and speedily integrate with a wide variety of databases.

Integrate Python into Microsoft Excel. Use Excel as your user-facing front-end with calculations, business logic and data access powered by Python. Works with all 3rd party and open source Python packages. No need to write any VBA!

Python-based ecosystem of open-source software for mathematics, science, and engineering. It contains modules for optimization, linear algebra, integration, interpolation, special functions, FFT, signal and image processing, ODE solvers and other tasks common in science and engineering.

Dataform helps you manage all data processes in your cloud data warehouse. Publish tables, write data tests and automate complex SQL workflows in a few minutes, so you can spend more time on analytics and less time managing infrastructure.

It is the collaboration of Apache Spark and Python. it is a Python API for Spark that lets you harness the simplicity of Python and the power of Apache Spark in order to tame Big Data.

It enable users to ingest, blend, cleanse and prepare diverse data from any source. With visual tools to eliminate coding and complexity, It puts the best quality data at the fingertips of IT and the business.

It is a versatile tool that supports a variety of workloads. It is composed of two parts: Dynamic task scheduling optimized for computation. This is similar to Airflow, Luigi, Celery, or Make, but optimized for interactive computational workloads. Big Data collections like parallel arrays, dataframes, and lists that extend common interfaces like NumPy, Pandas, or Python iterators to larger-than-memory or distributed environments. These parallel collections run on top of dynamic task schedulers.

It is a free and open-source data analytics, reporting and integration platform. KNIME integrates various components for machine learning and data mining through its modular data pipelining concept.

An end-to-end data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps.

It is the leader in data virtualization providing data access, data governance and data delivery capabilities across the broadest range of enterprise, cloud, big data, and unstructured data sources without moving the data from their original repositories.