Showing 106 open source projects for "deep learning with python"

View related business solutions
  • The CI/CD Platform built for Mobile DevOps Icon
    The CI/CD Platform built for Mobile DevOps

    For mobile app developers interested in a powerful CI/CD platform for mobile app development and mobile DevOps

    Save time, money, and developer frustration with fast, flexible, and scalable mobile CI/CD that just works. Whether you swear by native or would rather go cross-platform, we have you covered. From Swift to Objective-C, Java to Kotlin, as well as Xamarin, Cordova, Ionic, React Native, and Flutter: Whatever you choose, we will automatically configure your initial workflows and have you building in minutes.
    Learn More
  • The CRM you will want to use every day Icon
    The CRM you will want to use every day

    With CRM, Sales, and Marketing Automation in one, Act! gives you everything you need for happier clients, more revenue, and less stress.

    Act! Premium is perfect for small and midsize businesses looking to market better, sell more, and create customers for life. With unparalleled flexibility and freedom of choice, Act! Premium accommodates the unique ways you do business. Whether it’s customizations to fit your specific business or industry processes or your preferences for deployment and access, the possibilities with Act! Premium are limitless.
    Learn More
  • 1
    Deep Lake

    Deep Lake

    Data Lake for Deep Learning. Build, manage, and query datasets

    Deep Lake (formerly known as Activeloop Hub) is a data lake for deep learning applications. Our open-source dataset format is optimized for rapid streaming and querying of data while training models at scale, and it includes a simple API for creating, storing, and collaborating on AI datasets of any size. It can be deployed locally or in the cloud, and it enables you to store all of your data in one place, ranging from simple annotations to large videos.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    PyTorch-Tutorial-2nd

    PyTorch-Tutorial-2nd

    CV, NLP, LLM project applications, and advanced engineering deployment

    PyTorch-Tutorial-2nd is an open-source educational repository that provides structured tutorials for learning deep learning with the PyTorch framework. The project serves as a practical companion to a second edition of a PyTorch learning guide and is designed to help learners understand neural network concepts through hands-on coding examples. The repository covers a wide range of topics including tensor operations, neural network construction, model training workflows, and optimization strategies. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Ludwig AI

    Ludwig AI

    Low-code framework for building custom LLMs, neural networks

    Declarative deep learning framework built for scale and efficiency. Ludwig is a low-code framework for building custom AI models like LLMs and other deep neural networks. Declarative YAML configuration file is all you need to train a state-of-the-art LLM on your data. Support for multi-task and multi-modality learning. Comprehensive config validation detects invalid parameter combinations and prevents runtime failures.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Transform months of data modeling and coding into days. Icon
    Transform months of data modeling and coding into days.

    Automatically generate, document, and govern your entire data architecture.

    Efficiently model your business and data models, and generate code for your data pipelines, data lakehouse, and analytical applications
    Learn More
  • 5
    Anomaly Detection Learning Resources

    Anomaly Detection Learning Resources

    Anomaly detection related books, papers, videos, and toolboxes

    ...It includes materials covering a wide range of anomaly detection domains, including time series data, graph data, tabular datasets, and real-time monitoring systems. By compiling resources from multiple programming ecosystems such as Python, R, and other machine learning platforms, the repository allows users to discover both research papers and practical implementations.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    bitsandbytes

    bitsandbytes

    Accessible large language models via k-bit quantization for PyTorch

    ...The library has become widely used in machine learning pipelines that rely on parameter-efficient training techniques and low-precision inference.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Tongyi DeepResearch

    Tongyi DeepResearch

    Tongyi Deep Research, the Leading Open-source Deep Research Agent

    DeepResearch (Tongyi DeepResearch) is an open-source “deep research agent” developed by Alibaba’s Tongyi Lab designed for long-horizon, information-seeking tasks. It’s built to act like a research agent: synthesizing, reasoning, retrieving information via the web and documents, and backing its outputs with evidence. The model is about 30.5 billion parameters in size, though at any given token only ~3.3B parameters are active.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    NBA Sports Betting Machine Learning

    NBA Sports Betting Machine Learning

    NBA sports betting using machine learning

    NBA-Machine-Learning-Sports-Betting is an open-source Python project that applies machine learning techniques to predict outcomes of National Basketball Association games for analytical and betting-related research. The system gathers historical team statistics and game data spanning multiple seasons, beginning with the 2007–2008 NBA season and continuing through the present.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    GLM-4

    GLM-4

    GLM-4 series: Open Multilingual Multimodal Chat LMs

    ...The GLM-4-32B-0414 models are trained on ~15T high-quality data (including substantial synthetic reasoning data), then post-trained with preference alignment, rejection sampling, and reinforcement learning to improve instruction following, coding, function calling, and agent-style behaviors. The GLM-Z1-32B-0414 line adds deeper mathematical, coding, and logical reasoning via extended reinforcement learning and pairwise ranking feedback, while GLM-Z1-Rumination-32B-0414 introduces a “rumination” mode that performs longer, tool-using deep research for complex, open-ended tasks. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Powering the next decade of business messaging | Twilio MessagingX Icon
    Powering the next decade of business messaging | Twilio MessagingX

    For organizations interested programmable APIs built on a scalable business messaging platform

    Build unique experiences across SMS, MMS, Facebook Messenger, and WhatsApp – with our unified messaging APIs.
    Learn More
  • 10
    MedicalGPT

    MedicalGPT

    MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training

    MedicalGPT training medical GPT model with ChatGPT training pipeline, implementation of Pretraining, Supervised Finetuning, Reward Modeling and Reinforcement Learning. MedicalGPT trains large medical models, including secondary pre-training, supervised fine-tuning, reward modeling, and reinforcement learning training.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 11
    DeepSeek-V3

    DeepSeek-V3

    Powerful AI language model (MoE) optimized for efficiency/performance

    DeepSeek-V3 is a robust Mixture-of-Experts (MoE) language model developed by DeepSeek, featuring a total of 671 billion parameters, with 37 billion activated per token. It employs Multi-head Latent Attention (MLA) and the DeepSeekMoE architecture to enhance computational efficiency. The model introduces an auxiliary-loss-free load balancing strategy and a multi-token prediction training objective to boost performance. Trained on 14.8 trillion diverse, high-quality tokens, DeepSeek-V3...
    Downloads: 158 This Week
    Last Update:
    See Project
  • 12
    H2O LLM Studio

    H2O LLM Studio

    Framework and no-code GUI for fine-tuning LLMs

    Welcome to H2O LLM Studio, a framework and no-code GUI designed for fine-tuning state-of-the-art large language models (LLMs). You can also use H2O LLM Studio with the command line interface (CLI) and specify the configuration file that contains all the experiment parameters. To finetune using H2O LLM Studio with CLI, activate the pipenv environment by running make shell. With H2O LLM Studio, training your large language model is easy and intuitive. First, upload your dataset and then start...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    ML Retreat

    ML Retreat

    Machine Learning Journal for Intermediate to Advanced Topics

    ML Retreat is an open-source learning repository that serves as a structured journal documenting advanced topics in machine learning and artificial intelligence. The project compiles detailed notes, technical explanations, and curated resources that guide readers through complex concepts across modern AI research. Rather than functioning as a traditional tutorial series, the repository is organized as a learning journey that progressively explores increasingly advanced subjects. Topics...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    DeepSeek R1

    DeepSeek R1

    Open-source, high-performance AI model with advanced reasoning

    DeepSeek-R1 is an open-source large language model developed by DeepSeek, designed to excel in complex reasoning tasks across domains such as mathematics, coding, and language. DeepSeek R1 offers unrestricted access for both commercial and academic use. The model employs a Mixture of Experts (MoE) architecture, comprising 671 billion total parameters with 37 billion active parameters per token, and supports a context length of up to 128,000 tokens. DeepSeek-R1's training regimen uniquely...
    Downloads: 73 This Week
    Last Update:
    See Project
  • 15
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    MLC LLM is a machine learning compiler and deployment framework designed to enable efficient execution of large language models across a wide range of hardware platforms. The project focuses on compiling models into optimized runtimes that can run natively on devices such as GPUs, mobile processors, browsers, and edge hardware. By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. The system...
    Downloads: 30 This Week
    Last Update:
    See Project
  • 16
    FastDeploy

    FastDeploy

    High-performance Inference and Deployment Toolkit for LLMs and VLMs

    FastDeploy is an open-source inference and deployment toolkit designed to simplify the process of running and serving deep learning models across a wide range of hardware platforms. Developed within the PaddlePaddle ecosystem, the toolkit focuses on providing high-performance deployment capabilities for modern AI models including large language models and vision-language systems. The platform enables developers to deploy trained models quickly using optimized inference pipelines that support GPUs, specialized AI accelerators, and other hardware architectures. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    SWIFT LLM

    SWIFT LLM

    Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs

    SWIFT LLM is a comprehensive framework developed within the ModelScope ecosystem for training, fine-tuning, evaluating, and deploying large language models and multimodal models. The platform provides a full machine learning pipeline that supports tasks ranging from model pre-training to reinforcement learning alignment techniques. It integrates with popular inference engines such as vLLM and LMDeploy to accelerate deployment and runtime performance. The framework also includes support for...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    marqo

    marqo

    Tensor search for humans

    ...Marqo is a versatile and robust search and analytics engine that can be integrated into any website or application. Due to horizontal scalability, Marqo provides lightning-fast query times, even with millions of documents. Marqo helps you configure deep-learning models like CLIP to pull semantic meaning from images. It can seamlessly handle image-to-image, image-to-text and text-to-image search and analytics. Marqo adapts and stores your data in a fully schemaless manner. It combines tensor search with a query DSL that provides efficient pre-filtering. Tensor search allows you to go beyond keyword matching and search based on the meaning of text, images and other unstructured data. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Happy-LLM

    Happy-LLM

    Large Language Model Principles and Practice Tutorial from Scratch

    Happy-LLM is an open-source educational project created by the Datawhale AI community that provides a structured and comprehensive tutorial for understanding and building large language models from scratch. The project guides learners through the entire conceptual and practical pipeline of modern LLM development, starting with foundational natural language processing concepts and gradually progressing to advanced architectures and training techniques. It explains the Transformer...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    TTRL

    TTRL

    Test-Time Reinforcement Learning

    TTRL is an open-source framework for test-time reinforcement learning in large language models, with a particular focus on reasoning tasks where ground-truth labels are not available during inference. The project addresses the problem of how to generate useful reward signals from unlabeled test-time data, and its central insight is that common test-time scaling practices such as majority voting can be repurposed into reward estimates for online reinforcement learning. This makes the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    AIDE ML

    AIDE ML

    AI-Driven Exploration in the Space of Code

    ...The system repeatedly improves its generated code by exploring different implementation paths and selecting the best-performing solutions. AIDE ML is packaged as a Python toolkit with built-in utilities such as command-line tools, configuration presets, and visualization interfaces that allow researchers to observe how the search process evolves. The framework is designed for experimentation and academic research into automated programming and machine learning optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    llms-from-scratch-cn

    llms-from-scratch-cn

    Build a large language model from 0 only with Python foundation

    ...Through a collection of notebooks, code examples, and translated learning materials, users can explore how to implement components such as multi-head attention, data loaders, and training pipelines using Python and PyTorch.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    PRIME

    PRIME

    Scalable RL solution for advanced reasoning of language models

    PRIME is an open-source reinforcement learning framework designed to improve the reasoning capabilities of large language models through process-level rewards rather than relying only on final outputs. The system introduces the concept of process reinforcement through implicit rewards, allowing models to receive feedback on intermediate reasoning steps instead of evaluating only the final answer. This approach helps models learn better reasoning strategies and encourages them to generate...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    GLM-V

    GLM-V

    GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning

    GLM-V is an open-source vision-language model (VLM) series from ZhipuAI that extends the GLM foundation models into multimodal reasoning and perception. The repository provides both GLM-4.5V and GLM-4.1V models, designed to advance beyond basic perception toward higher-level reasoning, long-context understanding, and agent-based applications. GLM-4.5V builds on the flagship GLM-4.5-Air foundation (106B parameters, 12B active), achieving state-of-the-art results on 42 benchmarks across image,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    SwanLab

    SwanLab

    An open-source, modern-design AI training tracking and visualization

    SwanLab is an open-source experiment tracking and visualization platform designed to help machine learning engineers monitor, compare, and analyze the training of artificial intelligence models. The tool records training metrics, hyperparameters, model outputs, and experiment configurations so that developers can easily understand how different experiments perform over time. It provides a modern user interface for visualizing results, enabling teams to compare runs, track model performance...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB