8 projects for "parallel computing" with 2 filters applied:

  • Tremendous is the global payouts platform for businesses sending gift cards and money at scale. Icon
    Tremendous is the global payouts platform for businesses sending gift cards and money at scale.

    Getting started is simple: add a funding method and place your first order in minutes.

    Trusted by 20,000+ leading organizations, Tremendous has delivered billions of rewards and enables businesses to reach recipients across 230+ countries and regions. Recipients have 2,500+ payout options to choose from, including gift cards, prepaid cards, cash transfers, and charitable donations.
    Learn More
  • QA Wolf | We Write, Run and Maintain Tests Icon
    QA Wolf | We Write, Run and Maintain Tests

    For developer teams searching for a testing software

    QA Wolf is an AI-native service that delivers 80% automated E2E test coverage for web & mobile apps in weeks not years.
    Learn More
  • 1
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Xtuner

    Xtuner

    A Next-Generation Training Engine Built for Ultra-Large MoE Models

    Xtuner is a large-scale training engine designed for efficient training and fine-tuning of modern large language models, particularly mixture-of-experts architectures. The framework focuses on enabling scalable training for extremely large models while maintaining efficiency across distributed computing environments. Unlike traditional 3D parallel training strategies, XTuner introduces optimized parallelism techniques that simplify scaling and reduce system complexity when training massive models. The engine supports training models with hundreds of billions of parameters and enables long-context training with sequence lengths reaching tens of thousands of tokens. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Mars Framework

    Mars Framework

    Mars is a tensor-based unified framework for large-scale data

    Mars is a distributed computing framework designed to scale scientific computing and data science workloads across large clusters while preserving the familiar programming interfaces of common Python libraries. The project provides a tensor-based execution model that extends the capabilities of tools such as NumPy, pandas, and scikit-learn so that large datasets can be processed in parallel without rewriting code for distributed environments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    Genetic Programming in OpenCL is a parallel implementation of genetic programming targeted at heterogeneous devices, such as CPU and GPU. It is written in OpenCL, an open standard for portable parallel programming across many computing platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Dominate AI Search Results Icon
    Dominate AI Search Results

    Generative Al is shaping brand discovery. AthenaHQ ensures your brand leads the conversation.

    AthenaHQ is a cutting-edge platform for Generative Engine Optimization (GEO), designed to help brands optimize their visibility and performance across AI-driven search platforms like ChatGPT, Google AI, and more.
    Learn More
  • 5
    LARA

    LARA

    Lightweight Architecture for boundedly Rational Agents

    For the purpose of policy simulation in coupled social-ecological systems (e.g. energy supply), a credible modelling of actors – especially citizens – and their decision processes is needed. This requires a framework capable of handling high numbers of heterogeneous agents (several hundreds of thousands). LARA (Lightweight Architecture for boundedly Rational Agents) meets these requirements and fills the gap between frameworks without built-in psychological foundations and full-fledged...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    MapPSO
    MapPSO is a tool for Ontology Alignment, which uses Discrete Particle Swarm Optimisation. A particle swarm is used to search for the optimal alignment. The algorithm is massively parallel and adapts naturally on parallel architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Platform for parallel computation in the Amazon cloud, including machine learning ensembles written in R for computational biology and other areas of scientific research. Home to MR-Tandem, a hadoop-enabled fork of X!Tandem peptide search engine.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    GAP (Grid Agents Platform) Toolkit for Modeling and Simulation of Mobile Agents in Grid Environments. GAP is an abstraction over GridSim, a Grid Simulation Toolkit for Resource Modelling and Application Scheduling for Parallel and Distributed Computing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB