Showing 2 open source projects for "llama"

View related business solutions
  • Connect with customers in one app Icon
    Connect with customers in one app

    Businesses of all sizes seeking an AI-enhanced, all-in-one communication platform to unify voice, video, and messaging for improved team collaboration

    Dialpad Connect is an AI-powered unified communications platform that combines voice, video, and messaging to enhance team collaboration and customer interactions. It features real-time call transcription, automated call summaries, and AI-generated action items to help users stay focused during conversations. The platform integrates seamlessly with popular business apps like Salesforce, Zendesk, Microsoft Teams, and Google Workspace to streamline workflows. Designed for businesses of all sizes, Dialpad Connect delivers enterprise-grade reliability with 100% uptime SLA and robust disaster recovery. Security and privacy are core priorities, meeting standards like GDPR, HIPAA, and SOC 2 compliance. Dialpad Connect helps companies elevate customer experiences while boosting team productivity.
    Learn More
  • Composable, Open Source Payments Platform Icon
    Composable, Open Source Payments Platform

    Build or enhance your payments stack, while maintaining control with an open-source, full-stack and modular infrastructure.

    Juspay's Payments Orchestration Platform offers a comprehensive product suite for businesses, including open-source payment orchestration, global payouts, seamless authentication, payment tokenization, fraud & risk management, end-to-end reconciliation, unified payment analytics & more. The company’s offerings also include end-to-end white label payment gateway solutions & real-time payments infrastructure for banks. These solutions help businesses achieve superior conversion rates, reduce fraud, optimize costs, and deliver seamless customer experiences at scale.
    Learn More
  • 1
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    ...Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. It also detects the user’s shell environment automatically, allowing it to generate commands tailored to shells such as Bash, Zsh, or PowerShell.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 2
    LLaMA.go

    LLaMA.go

    llama.go is like llama.cpp in pure Golang

    ...The code of the project is based on the legendary ggml.cpp framework of Georgi Gerganov written in C++ with the same attitude to performance and elegance. Both models store FP32 weights, so you'll needs at least 32Gb of RAM (not VRAM or GPU RAM) for LLaMA-7B. Double to 64Gb for LLaMA-13B.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB