Showing 20 open source projects for "parallel computing"

View related business solutions
  • Remote Network Monitoring and Management for an IoT World Icon
    Remote Network Monitoring and Management for an IoT World

    The Only RMM Solution You Need

    Domotz is the premier Remote Network Monitoring and Management platform for IoT. We offer powerful network management software for MSP's, Integrators, Security Professionals, and Business Owners. Domotz enables the complete solution to cost-effectively manage and monitor your customers’ networks with plug and play setup, a friendly UX, and a comprehensive feature set, accessible from any desktop browser or mobile device. Utilize one interface to manage multiple networks at multiple locations anywhere in the World. One person can deploy remote monitoring and management in less than 15 minutes.
    Sign Up for Free
  • No-Nonsense Code-to-Cloud Security for Devs | Aikido Icon
    No-Nonsense Code-to-Cloud Security for Devs | Aikido

    Connect your GitHub, GitLab, Bitbucket or Azure DevOps account to start scanning your repos for free.

    Aikido provides a unified security platform for developers, combining 12 powerful scans like SAST, DAST, and CSPM. AI-driven AutoFix and AutoTriage streamline vulnerability management, while runtime protection blocks attacks.
    Learn More
  • 1
    The Julia Programming Language

    The Julia Programming Language

    High-level, high-performance dynamic language for technical computing

    Julia is a fast, open source high-performance dynamic language for technical computing. It can be used for data visualization and plotting, deep learning, machine learning, scientific computing, parallel computing and so much more. Having a high level syntax, Julia is easy to use for programmers of every level and background. Julia has more than 2,800 community-registered packages including various mathematical libraries, data manipulation tools, and packages for general purpose computing. ...
    Downloads: 15 This Week
    Last Update:
    See Project
  • 2
    ncnn

    ncnn

    High-performance neural network inference framework for mobile

    ncnn is a high-performance neural network inference computing framework designed specifically for mobile platforms. It brings artificial intelligence right at your fingertips with no third-party dependencies, and speeds faster than all other known open source frameworks for mobile phone cpu. ncnn allows developers to easily deploy deep learning algorithm models to the mobile platform and create intelligent APPs. It is cross-platform and supports most commonly used CNN networks, including...
    Downloads: 85 This Week
    Last Update:
    See Project
  • 3
    nndeploy

    nndeploy

    An Easy-to-Use and High-Performance AI Deployment Framework

    ...The framework focuses on making it easier to transform trained AI models into production-ready applications that can run efficiently on desktops, mobile devices, servers, and edge computing hardware. Developers can use visual workflows to design and configure AI processing pipelines by connecting modular nodes that represent different stages of the inference process. The system supports multiple inference engines and hardware accelerators, allowing the same AI workflow to run on different platforms without significant modifications. nndeploy also includes performance optimization techniques such as parallel execution, memory reuse, and hardware-accelerated operations to improve inference speed.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 4
    ROOT

    ROOT

    Analyzing, storing and visualizing big data, scientifically

    ...It is reliable, performant and well supported, easy to use and obtain, and strives to maximize the quantity and impact of scientific results obtained per unit cost, both of human effort and computing resources. ROOT provides a very efficient storage system for data models, that demonstrated to scale at the Large Hadron Collider experiments: Exabytes of scientific data are written in columnar ROOT format. ROOT comes with histogramming capabilities in an arbitrary number of dimensions, curve fitting, statistical modeling, and minimization, to allow the easy setup of a data analysis system that can query and process the data interactively or in batch mode, as well as a general parallel processing framework, RDataFrame, that can considerably speed up an analysis.
    Downloads: 14 This Week
    Last Update:
    See Project
  • One Unified Time Tracking Software For Projects, Billing, Pay and Compliance Icon
    One Unified Time Tracking Software For Projects, Billing, Pay and Compliance

    For companies of all sizes looking for a Time Tracking software

    Replicon's time-tracking platform is scalable and configurable to support the diverse needs of small, mid & large businesses with a remote and globally distributed workforce. Replicon’s Time Tracking is a cloud-based, enterprise-grade solution that tracks employee time across projects, tasks, presence, and absence to facilitate client billing, project costing, and compliant payroll processing. The scalable and configurable platform offers seamless integration with common business technology stacks, such as ERP, CRM, Accounting, and payroll solutions. With AI-powered time capture, mobile apps, and labor compliance as a service, Replicon makes time tracking hassle-free.
    Learn More
  • 5
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    ArrayFire

    ArrayFire

    ArrayFire, a general purpose GPU library

    ArrayFire is a general-purpose tensor library that simplifies the process of software development for the parallel architectures found in CPUs, GPUs, and other hardware acceleration devices. The library serves users in every technical computing market. Data structures in ArrayFire are smartly managed to avoid costly memory transfers and to take advantage of each performance feature provided by the underlying hardware. The community of ArrayFire developers invites you to build with us if you're interested and able to write top performing tensor functions. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    TensorRT

    TensorRT

    C++ library for high performance inference on NVIDIA GPUs

    ...With TensorRT, you can optimize neural network models trained in all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded, or automotive product platforms. TensorRT is built on CUDA®, NVIDIA’s parallel programming model, and enables you to optimize inference leveraging libraries, development tools, and technologies in CUDA-X™ for artificial intelligence, autonomous machines, high-performance computing, and graphics. With new NVIDIA Ampere Architecture GPUs, TensorRT also leverages sparse tensor cores providing an additional performance boost.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 8
    Xtuner

    Xtuner

    A Next-Generation Training Engine Built for Ultra-Large MoE Models

    Xtuner is a large-scale training engine designed for efficient training and fine-tuning of modern large language models, particularly mixture-of-experts architectures. The framework focuses on enabling scalable training for extremely large models while maintaining efficiency across distributed computing environments. Unlike traditional 3D parallel training strategies, XTuner introduces optimized parallelism techniques that simplify scaling and reduce system complexity when training massive models. The engine supports training models with hundreds of billions of parameters and enables long-context training with sequence lengths reaching tens of thousands of tokens. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Ray

    Ray

    A unified framework for scalable computing

    Modern workloads like deep learning and hyperparameter tuning are compute-intensive and require distributed or parallel execution. Ray makes it effortless to parallelize single machine code — go from a single CPU to multi-core, multi-GPU or multi-node with minimal code changes. Accelerate your PyTorch and Tensorflow workload with a more resource-efficient and flexible distributed execution framework powered by Ray. Accelerate your hyperparameter search workloads with Ray Tune. Find the best...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Diagnose and Resolve IT Issues in Real Time Icon
    Diagnose and Resolve IT Issues in Real Time

    Engage your employees and agents more efficiently with ScreenMeet as a seamless extension of your existing IT Service Delivery Platform.

    ScreenMeet’s unique combination of video calling, screen share, and remote desktop functionality lets you quickly diagnose hardware and software issues with no frustration.
    Learn More
  • 10
    Mars Framework

    Mars Framework

    Mars is a tensor-based unified framework for large-scale data

    Mars is a distributed computing framework designed to scale scientific computing and data science workloads across large clusters while preserving the familiar programming interfaces of common Python libraries. The project provides a tensor-based execution model that extends the capabilities of tools such as NumPy, pandas, and scikit-learn so that large datasets can be processed in parallel without rewriting code for distributed environments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    YOLO ROS

    YOLO ROS

    YOLO ROS: Real-Time Object Detection for ROS

    ...Darknet on the CPU is fast (approximately 1.5 seconds on an Intel Core i7-6700HQ CPU @ 2.60GHz × 8) but it's like 500 times faster on GPU! You'll have to have an Nvidia GPU and you'll have to install CUDA. The CMakeLists.txt file automatically detects if you have CUDA installed or not. CUDA is a parallel computing platform and application programming interface (API) model created by Nvidia.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    aCompute

    aCompute

    Aims to enable researcher to tap in to mobile computing capability

    This is a software agent based computing program that will enable researchers and other users to tap in computing power of machine available by sharing work load on the fly with zero configuration on network & resources A self organizing agent program that will understand network and its resource. where as the only job left to researcher is to split up jobs in several chunks of programs either parallel or sequential jobs and go issue the job (A visual Modeler or Scripting support need to be yet designed) Software agents will automatically manage the rest or resource management, sharing , cloning of tasks etc. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Genetic Programming in OpenCL is a parallel implementation of genetic programming targeted at heterogeneous devices, such as CPU and GPU. It is written in OpenCL, an open standard for portable parallel programming across many computing platforms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    LARA

    LARA

    Lightweight Architecture for boundedly Rational Agents

    For the purpose of policy simulation in coupled social-ecological systems (e.g. energy supply), a credible modelling of actors – especially citizens – and their decision processes is needed. This requires a framework capable of handling high numbers of heterogeneous agents (several hundreds of thousands). LARA (Lightweight Architecture for boundedly Rational Agents) meets these requirements and fills the gap between frameworks without built-in psychological foundations and full-fledged...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    MapPSO
    MapPSO is a tool for Ontology Alignment, which uses Discrete Particle Swarm Optimisation. A particle swarm is used to search for the optimal alignment. The algorithm is massively parallel and adapts naturally on parallel architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Platform for parallel computation in the Amazon cloud, including machine learning ensembles written in R for computational biology and other areas of scientific research. Home to MR-Tandem, a hadoop-enabled fork of X!Tandem peptide search engine.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Simdist lets you harness the power of cluster computing without any knowledge of parallel libraries such as MPI, and with no restrictions on programming language. Primarily targeted at evolutionary computing and similar master-slave configurations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    SPADES (System for Parallel Agent Discrete Event Simulation) is a middleware system for the creation of artificial intelligence agent-based simulations. SPADES supports robust distribution across machines and tracks the thinking time of agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    pySPACE

    pySPACE

    Signal Processing and Classification Environment in Python using YAML

    pySPACE is a modular software for processing of large data streams that has been specifically designed to enable distributed execution and empirical evaluation of signal processing chains. Various signal processing algorithms (so called nodes) are available within the software, from finite impulse response filters over data-dependent spatial filters (e.g. CSP, xDAWN) to established classifiers (e.g. SVM, LDA). pySPACE incorporates the concept of node and node chains of the MDP framework. Due...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    GAP (Grid Agents Platform) Toolkit for Modeling and Simulation of Mobile Agents in Grid Environments. GAP is an abstraction over GridSim, a Grid Simulation Toolkit for Resource Modelling and Application Scheduling for Parallel and Distributed Computing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB