Description
An interactive parallelization framework which is especially useful in configuring data science workload distribution. Eg. supports openMIP, MPI runs on High Performance Clusters
Interactive Parallel Computing with IPython alternatives and similar packages
Based on the "Science and Data Analysis" category.
Alternatively, view Interactive Parallel Computing with IPython alternatives based on common mentions on social networks and blogs.
-
Pandas
Flexible and powerful data analysis / manipulation library for Python, providing labeled data structures similar to R data.frame objects, statistical functions, and much more -
NumPy
The fundamental package for scientific computing with Python. -
SymPy
A computer algebra system written in pure Python -
NetworkX
Network Analysis in Python -
Dask
Parallel computing with task scheduling -
statsmodels
Statsmodels: statistical modeling and econometrics in Python -
Numba
NumPy aware dynamic Python compiler using LLVM -
PyMC
Bayesian Modeling and Probabilistic Programming in Python -
Getting Started
PyGWalker: Turn your pandas dataframe into an interactive UI for visual analysis -
Biopython
Official git repository for Biopython (originally converted from CVS) -
astropy
Astronomy and astrophysics core library -
orange
๐ :bar_chart: :bulb: Orange: Interactive data analysis -
blaze
NumPy and Pandas interface to Big Data -
RDKit
The official sources for the RDKit library -
Cubes
[NOT MAINTAINED] Light-weight Python OLAP framework for multi-dimensional data analysis -
Open Mining
Business Intelligence (BI) in Python, OLAP -
#<Sawyer::Resource:0x00007f547e829e00>
A unified interface for distributed computing. Fugue executes SQL, Python, Pandas, and Polars code on Spark, Dask and Ray without any rewrites. -
bcbio-nextgen
Validated, scalable, community developed variant calling, RNA-seq and small RNA analysis -
NIPY
Workflows and interfaces for neuroimaging packages -
bcolz
A columnar data container that can be compressed. -
bccb
Incubator for useful bioinformatics code, primarily in Python and R -
Neupy
NeuPy is a Tensorflow based python library for prototyping and building neural networks -
Bubbles
[NOT MAINTAINED] Bubbles โ Python ETL framework -
harold
An open-source systems and controls toolbox for Python3 -
signac
Manage large and heterogeneous data spaces on the file system. -
LynxKite
The complete graph data science platform -
PatZilla
PatZilla is a modular patent information research platform and data integration toolkit with a modern user interface and access to multiple data sources. -
Kotori
A flexible data historian based on InfluxDB, Grafana, MQTT, and more. Free, open, simple. -
Terkin
Datalogger for MicroPython and CPython. -
dask-memusage
A low-impact profiler to figure out how much memory each task in Dask is using -
cclib
A library for parsing and interpreting the results of computational chemistry packages. -
ElasticBatch
Elasticsearch tool for easily collecting and batch inserting Python data and pandas DataFrames -
Open Babel
A chemical toolbox designed to speak the many languages of chemical data.
InfluxDB - Power Real-Time Data Analytics at Scale
* Code Quality Rankings and insights are calculated and provided by Lumnify.
They vary from L1 to L5 with "L5" being the highest.
Do you think we are missing an alternative of Interactive Parallel Computing with IPython or a related project?
README
Interactive Parallel Computing with IPython
IPython Parallel (ipyparallel
) is a Python package and collection of CLI scripts for controlling clusters of IPython processes, built on the Jupyter protocol.
IPython Parallel provides the following commands:
- ipcluster - start/stop/list clusters
- ipcontroller - start a controller
- ipengine - start an engine
Install
Install IPython Parallel:
pip install ipyparallel
This will install and enable the IPython Parallel extensions for Jupyter Notebook and (as of 7.0) Jupyter Lab 3.0.
Run
Start a cluster:
ipcluster start
Use it from Python:
import os
import ipyparallel as ipp
cluster = ipp.Cluster(n=4)
with cluster as rc:
ar = rc[:].apply_async(os.getpid)
pid_map = ar.get_dict()
See the docs for more info.