Showing 28 open source projects for "parallel computing"

View related business solutions
  • Jesta I.S. | Enterprise Software For Retail and Supply Chain Icon
    Jesta I.S. | Enterprise Software For Retail and Supply Chain

    Transition from fragmented entry-level or legacy systems to an enterprise suite.

    Unify your people and operations across all departments and channels. Discover end-to-end retail, wholesale, and supply chain management software suites designed to scale.
    Learn More
  • World class QA, 100% done-for-you Icon
    World class QA, 100% done-for-you

    For engineering teams in search of a solution to design, manage and maintain E2E tests for their apps

    MuukTest is a test automation service that combines our own proprietary, AI-powered software with expert QA services to help you achieve world class test automation at a fraction of the in-house costs.
    Learn More
  • 1
    Dask

    Dask

    Parallel computing with task scheduling

    Dask is a Python library for parallel and distributed computing, designed to scale analytics workloads from single machines to large clusters. It integrates with familiar tools like NumPy, Pandas, and scikit-learn while enabling execution across cores or nodes with minimal code changes. Dask excels at handling large datasets that don’t fit into memory and is widely used in data science, machine learning, and big data pipelines.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    PyOpenCL

    PyOpenCL

    OpenCL integration for Python, plus shiny features

    PyOpenCL is a Python wrapper for the OpenCL framework, providing seamless access to parallel computing on CPUs, GPUs, and other accelerators. It enables developers to harness the full power of heterogeneous computing directly from Python, combining Python’s ease of use with the performance benefits of OpenCL. PyOpenCL also includes convenient features for managing memory, compiling kernels, and interfacing with NumPy, making it a preferred choice in scientific computing, data analysis, and machine learning workflows that demand acceleration.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    CUDA Python

    CUDA Python

    Performance meets Productivity

    ...The project is designed to simplify GPU programming by offering Pythonic abstractions while still exposing the full power of CUDA for advanced users. It integrates tightly with the broader Python GPU ecosystem, including Numba for kernel compilation and CCCL for parallel primitives, allowing developers to write performant code without leaving Python. The toolkit also includes utilities for profiling, memory management, distributed computing, and numerical operations, making it suitable for scientific computing, AI, and data processing workloads.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    Lithops

    Lithops

    A multi-cloud framework for big data analytics

    Lithops is an open-source serverless computing framework that enables transparent execution of Python functions across multiple cloud providers and on-prem infrastructure. It abstracts cloud providers like IBM Cloud, AWS, Azure, and Google Cloud into a unified interface and turns your Python functions into scalable, event-driven workloads. Lithops is ideal for data processing, ML inference, and embarrassingly parallel workloads, giving you the power of FaaS (Function-as-a-Service) without vendor lock-in. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Kinetic Software - Epicor ERP Icon
    Kinetic Software - Epicor ERP

    Discrete, make-to-order and mixed-mode manufacturers who need a global cloud ERP solution

    Grow, thrive, and compete in a global marketplace with Kinetic—an industry-tailored, cognitive ERP that helps you work smarter and stay connected.
    Learn More
  • 5
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Xtuner

    Xtuner

    A Next-Generation Training Engine Built for Ultra-Large MoE Models

    Xtuner is a large-scale training engine designed for efficient training and fine-tuning of modern large language models, particularly mixture-of-experts architectures. The framework focuses on enabling scalable training for extremely large models while maintaining efficiency across distributed computing environments. Unlike traditional 3D parallel training strategies, XTuner introduces optimized parallelism techniques that simplify scaling and reduce system complexity when training massive models. The engine supports training models with hundreds of billions of parameters and enables long-context training with sequence lengths reaching tens of thousands of tokens. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Ray

    Ray

    A unified framework for scalable computing

    Modern workloads like deep learning and hyperparameter tuning are compute-intensive and require distributed or parallel execution. Ray makes it effortless to parallelize single machine code — go from a single CPU to multi-core, multi-GPU or multi-node with minimal code changes. Accelerate your PyTorch and Tensorflow workload with a more resource-efficient and flexible distributed execution framework powered by Ray. Accelerate your hyperparameter search workloads with Ray Tune. Find the best...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Mars Framework

    Mars Framework

    Mars is a tensor-based unified framework for large-scale data

    Mars is a distributed computing framework designed to scale scientific computing and data science workloads across large clusters while preserving the familiar programming interfaces of common Python libraries. The project provides a tensor-based execution model that extends the capabilities of tools such as NumPy, pandas, and scikit-learn so that large datasets can be processed in parallel without rewriting code for distributed environments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9

    dispy

    Distributed and Parallel Computing with/for Python.

    dispy is a generic and comprehensive, yet easy to use framework for creating and using compute clusters to execute computations in parallel across multiple processors in a single machine (SMP), among many machines in a cluster, grid or cloud. dispy is well suited for data parallel (SIMD) paradigm where a computation (Python function or standalone program) is evaluated with different (large) datasets independently. dispy supports public / private / hybrid cloud computing, fog / edge computing.
    Leader badge
    Downloads: 49 This Week
    Last Update:
    See Project
  • Regpack: All-in-One Online Registration and Payment Software Icon
    Regpack: All-in-One Online Registration and Payment Software

    For camps, courses, virtual classes, client billing, events, conferences, meetings, afterschool programs, educational travel, retreats

    Regpack is a powerful onboarding, registration, and payments platform trusted by thousands of organizations worldwide. Our mission is simple: to give you the tools to automate busywork, streamline your processes, and keep your focus where it belongs, on growing your programs and serving your clients.
    Learn More
  • 10
    PyCNN

    PyCNN

    Image Processing with Cellular Neural Networks in Python

    Image Processing with Cellular Neural Networks in Python. Cellular Neural Networks (CNN) are a parallel computing paradigm that was first proposed in 1988. Cellular neural networks are similar to neural networks, with the difference that communication is allowed only between neighboring units. Image Processing is one of its applications. CNN processors were designed to perform image processing; specifically, the original application of CNN processors was to perform real-time ultra-high frame-rate (>10,000 frame/s) processing unachievable by digital processors.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11

    SAT-Assembler

    Scalable and accurate targeted gene assembly for large-scale NGS data

    ...It recovers genes from gene families of particular interest to biologists with high coverage, low chimera rate, and extremely low memory usage compared with exiting gene assembly tools. Moreover, it is naturally compatible with parallel computing platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    aCompute

    aCompute

    Aims to enable researcher to tap in to mobile computing capability

    This is a software agent based computing program that will enable researchers and other users to tap in computing power of machine available by sharing work load on the fly with zero configuration on network & resources A self organizing agent program that will understand network and its resource. where as the only job left to researcher is to split up jobs in several chunks of programs either parallel or sequential jobs and go issue the job (A visual Modeler or Scripting support need to be yet designed) Software agents will automatically manage the rest or resource management, sharing , cloning of tasks etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    PuSSH
    PuSSH is Pythonic, Ubiquitous SSH, a Python wrapper/script that runs commands in parallel on clusters/ranges of linux/unix machines via SSH, ideally where SSH is configured to use Kerberos, RSA/DSA keys, or ssh-agent as to avoid password authentication.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    GXP is a parallel/distributed shell, plus a parallel task execution engine that runs your Makefile in parallel on distributed machines. Very easy to install (no need to compile. install it on YOUR machine and use it on ALL machines).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15

    corunner

    A utility to transport files and execute at remote in prarallel

    Are you maitaining many machines and bored with the tedious operation of logging to those machines one by one just for checking the machine's status? Or are you losing patient for your simple script that have to run command in those machines in sequence? The corunner is just for those senarios. It's designed to speed up execution in your machines. It can executes command in handreds of or thousands of machines concurrently. Provide it with your files and command and the machines' location,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    BriCS

    SaaS for running simulation models in the cloud

    The Bristol Cloud Service simulation runner is a cloud computing Software as a Service designed to enable users to quickly launch simulation code on Amazon AWS' Elastic Compute Cloud (EC2). BRiCS is written in Python using the Django framework and interacts with EC2 using the boto API. BriCS enables multiple simulation runs to be launched in parallel from a web browser. Model configuration (parameters) files are uploaded via the browser and results files are downloadable on completion. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    AWK~plus is the next generation script practice environment. The AWK Language specifications and a main extension of GNU GAWK. Combination of Dynamic and Static typing. Parallel computing that a lock is free, and is thread safe at a language level.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 18
    PyDSH is a remote administration tool, consisting of pydsh and pydcp. Pydsh allows you to run a command on multiple hosts in parallel over RSH, SSH or Telnet, OR manage your SSH public keys. The pydcp command allows copying files to/from multiple hosts.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Platform for parallel computation in the Amazon cloud, including machine learning ensembles written in R for computational biology and other areas of scientific research. Home to MR-Tandem, a hadoop-enabled fork of X!Tandem peptide search engine.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    cca-forum
    Cca-forum unifies the Common Component Architecture tools and tutorial. It includes the CCA specifications, the Ccaffeine framework for HPC, and related tools. These support multilanguage scientific and parallel computing.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    Distributed Parallel Programming for Python! This package builds on traditional Python by enabling users to write distributed, parallel programs based on MPI message passing primitives. General python objects can be messaged between processors. Ru
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    PyMW is a Python module for parallel master-worker computing in a variety of environments. With the PyMW module, users can write a single program that scales from multicore machines to global computing platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Pydusa is a package for parallel programming using Python. It contains a module for doing MPI programming in Python. We have added parallel solver packages such as Parallel SuperLU for solving sparse linear systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Python Integrated Parallel Programming EnviRonment (PIPPER), Python pre-parser that is designed to manage a pipeline, written in Python. It enables automated parallelization of loops. Think of it like OpenMP for Python, but it works in a computer cluster
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Simple Distributed Job Management Tools facilitates parallel, distributed execution of simple commands, on a network of UNIX-like machines. Components include a job dependency/exclusivity language and load-balanced remote execution facility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB