Showing 181 open source projects for "git:/git.code.sf.net/p/docfetcher/code"

View related business solutions
  • All Things Performance and Partner Marketing, All in One Place Icon
    All Things Performance and Partner Marketing, All in One Place

    Track calls, leads, and clicks without the manual work

    Automatically tie revenue back to campaigns, channels, publishers, and networks through marketing attribution. Spend less time juggling reports, and more time optimizing for growth by using a single operating solution for partner and performance marketing.
    Learn More
  • MaintainX is the world-leading mobile-first workflow management platform for industrial and frontline workers. Icon
    MaintainX is the world-leading mobile-first workflow management platform for industrial and frontline workers.

    Trusted by Operational Leaders Across the Globe

    Your day-to-day maintenance tasks, simplified. MaintainX eliminates the paperwork, so you can spend less time on your clipboard and more time getting things done.
    Learn More
  • 1
    AIConfig

    AIConfig

    AIConfig is a config-based framework to build generative AI apps

    AIConfig is an open-source framework designed to simplify the development and management of generative AI applications by separating AI logic from application code. The framework allows prompts, model configurations, and parameters to be stored as structured configuration files that can be version controlled and managed independently from the rest of the software system. This approach improves collaboration between developers, prompt engineers, and machine learning practitioners by turning prompt logic into a reusable and editable artifact. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    LIDA

    LIDA

    Automatic Generation of Visualizations and Infographics using LLMs

    LIDA is an open-source library developed to automate the process of creating data visualizations and infographics using large language models. The system treats visualizations as executable code and uses AI to generate, modify, and interpret that code in order to transform raw datasets into meaningful charts and graphical explanations. Instead of requiring users to manually explore datasets and write plotting scripts, LIDA analyzes the data and automatically proposes visualization goals and design ideas that highlight patterns and relationships. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Automated Interpretability

    Automated Interpretability

    Code for Language models can explain neurons in language models paper

    The automated-interpretability repository implements tools and pipelines for automatically generating, simulating, and scoring explanations of neuron (or latent feature) behavior in neural networks. Instead of relying purely on manual, ad hoc interpretability probing, this repo aims to scale interpretability by using algorithmic methods that produce candidate explanations and assess their quality. It includes a “neuron explainer” component that, given a target neuron or latent feature,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Qwen2.5-Coder

    Qwen2.5-Coder

    Qwen2.5-Coder is the code version of Qwen2.5, the large language model

    Qwen2.5-Coder, developed by QwenLM, is an advanced open-source code generation model designed for developers seeking powerful and diverse coding capabilities. It includes multiple model sizes—ranging from 0.5B to 32B parameters—providing solutions for a wide array of coding needs. The model supports over 92 programming languages and offers exceptional performance in generating code, debugging, and mathematical problem-solving.
    Downloads: 9 This Week
    Last Update:
    See Project
  • Enterprise Job Scheduling Software Icon
    Enterprise Job Scheduling Software

    Unify Enterprise Job Scheduling for Scale, Visibility, and Control

    Managing your sprawling data center and cloud with disparate native schedulers creates chaos. Achieve unparalleled control and efficiency over your entire IT environment with JAMS job orchestration tools. JAMS provides the singular, centralized platform required to overcome the complexities of disparate native schedulers. Automate, secure, and govern all your workloads, eliminating fragmented control, compliance risks, and operational bottlenecks. JAMS streamlines operations and ensures audit-ready history, transforming your enterprise automation with confidence and precision.
    Learn More
  • 5
    ToRA

    ToRA

    Tool-integrated Reasoning LLM Agents

    ...Instead of relying solely on text generation, the system dynamically invokes tools such as symbolic solvers or programming libraries when deeper computation is required. This approach allows the model to reason step by step in natural language and then execute precise calculations or code through tool calls, creating a hybrid reasoning workflow. The framework was designed to address known weaknesses of large language models in mathematical problem solving and formal reasoning tasks. Training data includes tool-use trajectories that teach the model when to reason verbally and when to delegate tasks to specialized tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DeepSeek LLM

    DeepSeek LLM

    DeepSeek LLM: Let there be answers

    The DeepSeek-LLM repository hosts the code, model files, evaluations, and documentation for DeepSeek’s LLM series (notably the 67B Chat variant). Its tagline is “Let there be answers.” The repo includes an “evaluation” folder (with results like math benchmark scores) and code artifacts (e.g. pre-commit config) that support model development and deployment. According to the evaluation files, DeepSeek LLM 67B Chat achieves strong performance on math benchmarks under both chain-of-thought (CoT) and tool-assisted reasoning modes. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 7
    Grok-1

    Grok-1

    Open-source, high-performance Mixture-of-Experts large language model

    ...In March 2024, xAI released Grok-1's model weights and architecture under the Apache 2.0 license, making them openly accessible to developers. The accompanying GitHub repository provides JAX example code for loading and running the model. Due to its substantial size, utilizing Grok-1 requires a machine with significant GPU memory. The repository's MoE layer implementation prioritizes correctness over efficiency, avoiding the need for custom kernels. This is a full repo snapshot ZIP file of the Grok-1 code.
    Leader badge
    Downloads: 33 This Week
    Last Update:
    See Project
  • 8
    Bard API

    Bard API

    The unofficial python package that returns response of Google Bard

    ...Therefore, I strongly discourage using it for any other purposes. If you have access to official PaLM-2 API, replace the provided response with the corresponding official code.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    MGIE

    MGIE

    Guiding Instruction-based Image Editing via Multimodal Large Language

    ...The project focuses on making edits explainable and controllable: the model interprets text guidance, reasons over image content, and outputs edits aligned with user intent. It’s positioned as an ICLR 2024 Spotlight work, with code and references that show how to connect language planning to concrete image operations. This bridges a gap between free-form prompts and precise edits by letting users describe “what” and “where” in everyday language. The repo includes instructions, examples, and links that situate MGIE within Apple’s broader line of multimodal research. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • The most advanced C and C++ source code analyzer Icon
    The most advanced C and C++ source code analyzer

    Combining the benefits of static and dynamic source code analysis to deliver the most advanced & exhaustive code verification tool.

    TrustInSoft Analyzer is a C and C++ source code analyzer powered by formal methods, mathematical & logical reasonings that allow for exhaustive analysis of source code. This analysis can be run without false positives or false negatives, so that every real bug in the code is found. Developers receive several benefits: a user-friendly graphical interface that directs developers to the root cause of bugs, and instant utility to expand the coverage of their existing tests. Unlike traditional source code analysis tools, TrustInSoft’s solution is not only the most comprehensive approach on the market but is also progressive, instantly deployable by developers, even if they lack experience with formal methods, from exhaustive analysis up to a functional proof that the software developed meets specifications.
    Learn More
  • 10
    Qwen2.5

    Qwen2.5

    Open source large language model by Alibaba

    ...The models are open-source under the Apache 2.0 license, with resources and documentation available on platforms like Hugging Face and ModelScope. This is a full ZIP snapshot of the Qwen2.5 code.
    Downloads: 21 This Week
    Last Update:
    See Project
  • 11
    Mixtral offloading

    Mixtral offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

    ...This approach takes advantage of the sparse activation properties of mixture-of-experts architectures, where only a subset of expert networks are used for each token during generation. By selectively loading and caching the required experts, the system avoids keeping the entire model in GPU memory at once. The repository includes notebooks and code examples that demonstrate how to run large language models on consumer hardware such as personal GPUs or cloud notebook environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    EvaDB

    EvaDB

    Database system for building simpler and faster AI-powered application

    Over the last decade, AI models have radically changed the world of natural language processing and computer vision. They are accurate on various tasks ranging from question answering to object tracking in videos. To use an AI model, the user needs to program against multiple low-level libraries, like PyTorch, Hugging Face, Open AI, etc. This tedious process often leads to a complex AI app that glues together these libraries to accomplish the given task. This programming complexity prevents...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Mr. Ranedeer

    Mr. Ranedeer

    GPT-4 AI Tutor Prompt for customizable personalized learning

    Unlock the potential of GPT-4 with Mr. Ranedeer AI Tutor, a customizable prompt that delivers personalized learning experiences for users with diverse needs and interests.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    AutoGPT.js

    AutoGPT.js

    Auto-GPT on the browser

    AutoGPT.js is an open-source project that brings autonomous AI agent capabilities similar to AutoGPT directly into the browser environment. The system allows users to run an AI agent capable of performing tasks such as generating code, searching the web, and interacting with files on the local computer. Unlike traditional AutoGPT implementations that require server infrastructure, AutoGPT.js is designed to run primarily in the browser, making it easier to deploy and experiment with autonomous agents. The platform uses web APIs and language model integrations to give the agent the ability to plan tasks, execute commands, and store short-term memory during operations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    ReplitLM

    ReplitLM

    Inference code and configs for the ReplitLM model family

    ReplitLM is a family of open-source language models developed by Replit for assisting with programming tasks such as code generation and completion. The project includes model checkpoints, configuration files, and example code that enable developers to run and experiment with the models locally or within machine learning frameworks. These models are designed specifically for coding workflows and are trained on large datasets of source code covering many programming languages and development environments. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Autolabel

    Autolabel

    Label, clean and enrich text datasets with LLMs

    Autolabel is a Python library to label, clean and enrich datasets with Large Language Models (LLMs). Autolabel data for NLP tasks such as classification, question-answering and named entity recognition, entity matching and more. Seamlessly use commercial and open-source LLMs from providers such as OpenAI, Anthropic, HuggingFace, Google and more.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Llama 2 Everywhere (L2E)

    Llama 2 Everywhere (L2E)

    Llama 2 Everywhere (L2E)

    Llama 2 Everywhere (L2E) is an open-source implementation of the LLaMA-2 large language model architecture designed to demonstrate how transformer-based language models can be executed with extremely minimal code. The project focuses on simplicity and educational clarity by implementing inference for LLaMA-style models in a compact C program rather than relying on large machine learning frameworks. Developers can train models using a Python training pipeline and then run inference using a lightweight C implementation that requires very few dependencies. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    mindflow

    mindflow

    AI-powered CLI git wrapper, boilerplate code generator, chat history

    I-powered CLI git wrapper, boilerplate code generator, chat history manager, and code search engine to streamline your dev workflow. The ChatGPT-powered swiss army knife for the modern developer! We provide an AI-powered CLI git wrapper, boilerplate code generator, code search engine, a conversation history manager, and much more! Configure the model used for generating responses by running mf config and selecting either GPT 3.5 Turbo (default) or GPT 4. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    LLaMA

    LLaMA

    Inference code for Llama models

    “Llama” is the repository from Meta (formerly Facebook/Meta Research) containing the inference code for LLaMA (Large Language Model Meta AI) models. It provides utilities to load pre-trained LLaMA model weights, run inference (text generation, chat, completions), and work with tokenizers. Tokenizer utilities, download scripts, shell helpers to fetch model weights with correct licensing/permissions. Includes example scripts for chat completions and text completions to show how to call the models in code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    LangChain Apps on Production with Jina

    LangChain Apps on Production with Jina

    Langchain Apps on Production with Jina & FastAPI

    Jina is an open-source framework for building scalable multi-modal AI apps on Production. LangChain is another open-source framework for building applications powered by LLMs. long-chain-serve helps you deploy your LangChain apps on Jina AI Cloud in a matter of seconds. You can benefit from the scalability and serverless architecture of the cloud without sacrificing the ease and convenience of local development. And if you prefer, you can also deploy your LangChain apps on your own...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    aigc

    aigc

    An e-book about the real-world application of LLM

    "Building Large Language Model Applications: Application Development and Architecture Design" is an open source e-book about the real-world application of LLM. It introduces the basics and applications of large language models, as well as how to build your own models. These include writing, developing, and managing prompts, exploring what the best large language models can bring, and pattern and architecture design for LLM application development.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    Repo of Tree of Thoughts (ToT)

    Repo of Tree of Thoughts (ToT)

    Implementation of "Tree of Thoughts

    Language models are increasingly being deployed for general problem-solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. To surmount these challenges, we introduce a new framework for language model inference, Tree of Thoughts (ToT), which generalizes over the popular Chain of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    aqueduct LLM

    aqueduct LLM

    Aqueduct allows you to run LLM and ML workloads on any infrastructure

    Aqueduct is an MLOps framework that allows you to define and deploy machine learning and LLM workloads on any cloud infrastructure. Aqueduct is an open-source MLOps framework that allows you to write code in vanilla Python, run that code on any cloud infrastructure you'd like to use, and gain visibility into the execution and performance of your models and predictions. Aqueduct's Python native API allows you to define ML tasks in regular Python code. You can connect Aqueduct to your existing cloud infrastructure (docs), and Aqueduct will seamlessly move your code from your laptop to the cloud or between different cloud infrastructure layers. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    unit-minions

    unit-minions

    AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA

    "AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA", including Llama (Alpaca LoRA) model, ChatGLM (ChatGLM Tuning) related Lora training. Training content: user story generation, test code generation, code-assisted generation, text to SQL, text generation code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ChatLLM Web

    ChatLLM Web

    Chat with LLM like Vicuna totally in your browser with WebGPU

    Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered By web-llm. To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions ≤ 112 are not supported. You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower. The first time you use the app, you will need to download the model. For the Vicuna-7b model that...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB