Search Results for "/storage/emulated/0/android/data/net.sourceforge.uiq3.fx603p/files" - Page 43

Showing 2237 open source projects for "/storage/emulated/0/android/data/net.sourceforge.uiq3.fx603p/files"

View related business solutions
  • The AI workplace management platform Icon
    The AI workplace management platform

    Plan smart spaces, connect teams, manage assets, and get insights with the leading AI-powered operating system for the built world.

    By combining AI workflows, predictive intelligence, and automated insights, OfficeSpace gives leaders a complete view of how their spaces are used and how people work. Facilities, IT, HR, and Real Estate teams use OfficeSpace to optimize space utilization, enhance employee experience, and reduce portfolio costs with precision.
    Learn More
  • Securden Privileged Account Manager Icon
    Securden Privileged Account Manager

    Unified Privileged Access Management

    Discover and manage administrator, service, and web app passwords, keys, and identities. Automate management with approval workflows. Centrally control, audit, monitor, and record all access to critical IT assets.
    Learn More
  • 1
    KoboldCpp

    KoboldCpp

    Run GGUF models easily with a UI or API. One File. Zero Install.

    KoboldCpp is an easy-to-use AI text-generation software for GGML and GGUF models, inspired by the original KoboldAI. It's a single self-contained distributable that builds off llama.cpp and adds many additional powerful features.
    Leader badge
    Downloads: 525 This Week
    Last Update:
    See Project
  • 2
    spacy-llm

    spacy-llm

    Integrating LLMs into structured NLP pipelines

    ...This package integrates Large Language Models (LLMs) into spaCy, featuring a modular system for fast prototyping and prompting, and turning unstructured responses into robust outputs for various NLP tasks, no training data required.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    TorchMetrics

    TorchMetrics

    Machine learning metrics for distributed, scalable PyTorch application

    TorchMetrics is a collection of 80+ PyTorch metrics implementations and an easy-to-use API to create custom metrics. Your data will always be placed on the same device as your metrics. You can log Metric objects directly in Lightning to reduce even more boilerplate. The module-based metrics contain internal metric states (similar to the parameters of the PyTorch module) that automate accumulation and synchronization across devices! Automatic accumulation over multiple batches. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    autoresearch-mlx

    autoresearch-mlx

    Apple Silicon (MLX) port of Karpathy's autoresearch

    ...The system is tailored for Apple hardware, leveraging unified memory and MLX capabilities to achieve efficient training on Mac devices. It includes a minimal and focused project structure consisting of data preparation utilities, a modifiable training file, and a program specification that governs the agent’s behavior. The framework logs experiment results and supports continuous iteration, enabling long-running optimization cycles that can reveal hardware-specific performance patterns.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Network Management Software and Tools for Businesses and Organizations | Auvik Networks Icon
    Network Management Software and Tools for Businesses and Organizations | Auvik Networks

    Mapping, inventory, config backup, and more.

    Reduce IT headaches and save time with a proven solution for automated network discovery, documentation, and performance monitoring. Choose Auvik because you'll see value in minutes, and stay with us to improve your IT for years to come.
    Learn More
  • 5
    TNT

    TNT

    A lightweight library for PyTorch training tools and utilities

    TNT is a lightweight training framework developed by Meta that simplifies the process of building and managing machine learning training loops using PyTorch. The project focuses on providing a flexible yet structured environment for implementing training pipelines without the complexity of large deep learning frameworks. It introduces modular abstractions that allow developers to organize training logic into reusable components such as trainers, evaluators, and callbacks. This design helps...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    The Hypersim Dataset

    The Hypersim Dataset

    Photorealistic Synthetic Dataset for Holistic Indoor Scene

    ...The dataset spans diverse furniture layouts, room types, and camera trajectories, enabling robust training for geometry, segmentation, and SLAM-adjacent tasks. Rendering pipelines and utilities allow researchers to reproduce sequences, generate novel views, or extract task-specific supervision. Because the data are perfectly labeled and controllable, Hypersim is well suited for pretraining and for studying domain transfer to real imagery. The repository acts as both a dataset index and a set of scripts for downloading, managing, and evaluating on standardized splits.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Autoskills

    Autoskills

    One command. Your entire AI skill stack. Installed

    The Autoskills project is a developer tool that automates the installation of AI agent skills based on a project’s technology stack. It operates through a simple command-line interface that scans configuration files such as package.json and build scripts to detect the frameworks, languages, and tools used in a project. Once the stack is identified, it automatically installs a curated set of AI skills tailored to those technologies, significantly reducing setup time for AI-assisted development environments. The system is designed to work across a wide range of ecosystems, including frontend, backend, mobile, cloud, and AI tooling stacks. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Khazix Skills

    Khazix Skills

    Digital Life Kazik Open Source AI Skills Collection

    ...It acts as a pipeline that analyzes a repository’s metadata, extracts relevant information such as README content and commit hashes, and converts it into a standardized skill format that can be integrated into agent ecosystems. The system emphasizes lifecycle management by embedding versioning, traceability, and metadata directly into generated skill files, allowing future updates and synchronization with the original repository. It also generates wrapper scripts that enable AI agents to interact with the underlying repository functionality without requiring deep manual integration. By enforcing a consistent schema, the project ensures interoperability between skills and simplifies deployment across environments. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    ModernBERT

    ModernBERT

    Bringing BERT into modernity via both architecture changes and scaling

    ModernBERT is an open-source research project that modernizes the classic BERT encoder architecture by incorporating recent advances in transformer design, training techniques, and efficiency improvements. The goal of the project is to bring BERT-style models up to date with the capabilities of modern large language models while preserving the strengths of bidirectional encoder architectures used for tasks such as classification, retrieval, and semantic search. ModernBERT introduces...
    Downloads: 1 This Week
    Last Update:
    See Project
  • The AI-powered unified PSA-RMM platform for modern MSPs. Icon
    The AI-powered unified PSA-RMM platform for modern MSPs.

    Trusted PSA-RMM partner of MSPs worldwide

    SuperOps.ai is the only PSA-RMM platform powered by intelligent automation and thoughtfully crafted for the new-age MSP. The platform also helps MSPs manage their projects, clients, and IT documents from a single place.
    Learn More
  • 10
    GitHub Agentic Workflows

    GitHub Agentic Workflows

    GitHub Agentic Workflows

    GitHub Agentic Workflows is an experimental CLI extension and framework for the gh GitHub CLI that lets developers author automation driven by natural language specifications instead of hand-written code, compiling those descriptions into GitHub Actions workflows that run AI agents (like Copilot, Claude Code, or Codex) on schedule or in response to repository events. By writing intent in markdown files, a developer can quickly generate .yml Actions workflows that perform tasks such as summarizing issues, automating triage, generating reports, or maintaining documentation, all without manually crafting YAML logic from scratch. The system emphasizes safety and guardrails, running agents in sandboxed environments with minimal permissions by default, and using “safe outputs” to constrain what the workflow can write back into the repository. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Open Model Zoo

    Open Model Zoo

    Pre-trained Deep Learning models and demos

    ...It includes hundreds of models covering object detection, classification, segmentation, pose estimation, speech recognition, text-to-speech, and more, many of which are already converted into formats optimized for inference on CPUs, GPUs, VPUs, and other accelerators supported by OpenVINO. In addition to model files, Open Model Zoo provides demo applications that show realistic usage patterns and help developers quickly prototype and understand inference pipelines in C++, Python, or via the OpenCV Graph API. Tools in the repository also help automate model downloads and other tasks, making it easier to incorporate these models into production systems or custom solutions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Core ML Stable Diffusion

    Core ML Stable Diffusion

    Stable Diffusion with Core ML on Apple Silicon

    ...StableDiffusion, a Swift package that developers can add to their Xcode projects as a dependency to deploy image generation capabilities in their apps. The Swift package relies on the Core ML model files generated by python_coreml_stable_diffusion. Hugging Face ran the conversion procedure on the following models and made the Core ML weights publicly available on the Hub. If you would like to convert a version of Stable Diffusion that is not already available on the Hub, please refer to the Converting Models to Core ML. Log in to or register for your Hugging Face account, generate a User Access Token and use this token to set up Hugging Face API access by running huggingface-cli login in a Terminal window.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    MCP Security Audit

    MCP Security Audit

    Powerful MCP (Model Context Protocol) Server that audits npm packages

    The MCP Security Audit Server is a powerful Model Context Protocol server that audits npm package dependencies for security vulnerabilities, integrating with remote npm registries for real-time security checks. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Langchainrb

    Langchainrb

    Build LLM-powered applications in Ruby

    LangchainRB is a Ruby implementation of LangChain, allowing developers to build AI-driven applications using large language models (LLMs) and knowledge graphs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Milvus Bootcamp

    Milvus Bootcamp

    Dealing with all unstructured data, such as reverse image search

    Milvus Bootcamp is a collection of tutorials, examples, and best practices for using Milvus, an open-source vector database designed for AI-powered similarity search and retrieval applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Flock

    Flock

    Flock is a workflow-based low-code platform for building chatbots

    Flock is a workflow-based low-code platform designed for building AI applications such as chatbots, retrieval-augmented generation systems, and multi-agent workflows. The platform uses a visual workflow architecture where different nodes represent processing steps such as input processing, model inference, retrieval operations, and tool execution. Developers can connect these nodes to create complex pipelines that orchestrate multiple language models and external services. Built on...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    Microsoft Learn MCP Server

    Microsoft Learn MCP Server

    Official Microsoft Learn MCP Server, powering LLMs and AI agents

    Microsoft Learn MCP Server is the official GitHub repository for the Microsoft Learn MCP (Model Context Protocol) Server, a service that implements the Model Context Protocol to provide AI assistants and tools with reliable, real-time access to Microsoft’s official documentation. Rather than relying on training data that may be outdated or incomplete, MCP servers let agents like GitHub Copilot, Claude, or other LLM-based tools search and pull context directly from up-to-date Microsoft Learn content, including Azure, .NET, and other tech docs. By connecting to the MCP endpoint, coding agents can answer questions, retrieve code examples, and offer best practices grounded in authoritative sources without requiring API keys or manual browser searches. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    GenAI Agents

    GenAI Agents

    Implementations for various Generative AI Agent techniques

    GenAI Agents is a large, tutorial-driven repository that teaches you how to design, build, and experiment with generative AI agents. It spans a spectrum from simple conversational bots and basic question-answering agents to complex multi-agent systems that coordinate on research, education, business workflows, and creative tasks. The implementations leverage modern frameworks such as LangChain, LangGraph, AutoGen, PydanticAI, CrewAI, and more, showing how each can be wired into realistic...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel. The interface mirrors the modern chat UX you’d expect—streaming responses, markdown, and a clean layout—so there’s no usability tradeoff to gain privacy. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    TensorFlow Serving

    TensorFlow Serving

    Serving system for machine learning models

    ...It deals with the inference aspect of machine learning, taking models after training and managing their lifetimes, providing clients with versioned access via a high-performance, reference-counted lookup table. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data. The easiest and most straight-forward way of using TensorFlow Serving is with Docker images. We highly recommend this route unless you have specific needs that are not addressed by running in a container. In order to serve a Tensorflow model, simply export a SavedModel from your Tensorflow program. SavedModel is a language-neutral, recoverable, hermetic serialization format that enables higher-level systems and tools to produce, consume, and transform TensorFlow models.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    LLMFarm

    LLMFarm

    llama and other large language models on iOS and MacOS offline

    LLMFarm is a framework designed to simplify the deployment, management, and utilization of large language models in local or self-hosted environments, focusing on accessibility and efficient resource usage. It enables users to run LLMs on personal hardware or private infrastructure, reducing dependency on external APIs and improving data privacy. The system typically provides a user-friendly interface for loading models, configuring inference parameters, and interacting with them through chat or task-based workflows. It emphasizes modularity, allowing users to integrate different models, backends, or tools depending on their needs and hardware capabilities. LLMFarm is particularly useful for developers and researchers experimenting with local AI systems, as it lowers the barrier to entry for running and testing models without extensive setup. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Cosmos-RL

    Cosmos-RL

    Cosmos-RL is a flexible and scalable Reinforcement Learning framework

    ...It provides a distributed training architecture that separates policy learning and environment rollout processes, enabling efficient and asynchronous reinforcement learning at scale. The framework supports multiple parallelism strategies, including tensor, pipeline, and data parallelism, allowing it to leverage large GPU clusters effectively. It is built with compatibility in mind, supporting popular model families such as LLaMA, Qwen, and diffusion-based world models, as well as integration with Hugging Face ecosystems. cosmos-rl also includes support for advanced RL algorithms, low-precision training, and fault-tolerant execution, making it suitable for large-scale production workloads.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    autoMate

    autoMate

    AI tool for automating desktop tasks via natural language input

    ...Unlike conventional RPA tools that require predefined workflows, autoMate dynamically adapts to tasks by making autonomous decisions based on the current interface state. autoMate emphasizes local execution, meaning all processing happens on the user’s machine to maintain privacy and data security.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Linfa

    Linfa

    A Rust machine learning framework

    linfa aims to provide a comprehensive toolkit to build Machine Learning applications with Rust. Kin in spirit to Python's scikit-learn, it focuses on common preprocessing tasks and classical ML algorithms for your everyday ML tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    CML

    CML

    Continuous Machine Learning | CI/CD for ML

    Continuous Machine Learning (CML) is an open-source CLI tool for implementing continuous integration & delivery (CI/CD) with a focus on MLOps. Use it to automate development workflows, including machine provisioning, model training and evaluation, comparing ML experiments across project history, and monitoring changing datasets. CML can help train and evaluate models, and then generate a visual report with results and metrics, automatically on every pull request.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB