Data Science Tools

View 126 business solutions
  • The full-stack observability platform that protects your dataLayer, tags and conversion data Icon
    The full-stack observability platform that protects your dataLayer, tags and conversion data

    Stop losing revenue to bad data today. and protect your marketing data with Code-Cube.io.

    Code-Cube.io detects issues instantly, alerts you in real time and helps you resolve them fast. No manual QA. No unreliable data. Just data you can trust and act on.
    Learn More
  • Outbound sales software Icon
    Outbound sales software

    Unified cloud-based platform for dialing, emailing, appointment scheduling, lead management and much more.

    Adversus is an outbound dialing solution that helps you streamline your call strategies, automate manual processes, and provide valuable insights to improve your outbound workflows and efficiency.
    Learn More
  • 1
    ML workspace

    ML workspace

    All-in-one web-based IDE specialized for machine learning

    All-in-one web-based development environment for machine learning. The ML workspace is an all-in-one web-based IDE specialized for machine learning and data science. It is simple to deploy and gets you started within minutes to productively built ML solutions on your own machines. This workspace is the ultimate tool for developers preloaded with a variety of popular data science libraries (e.g., Tensorflow, PyTorch, Keras, Sklearn) and dev tools (e.g., Jupyter, VS Code, Tensorboard) perfectly configured, optimized, and integrated. Usable as remote kernel (Jupyter) or remote machine (VS Code) via SSH. Easy to deploy on Mac, Linux, and Windows via Docker. Jupyter, JupyterLab, and Visual Studio Code web-based IDEs.By default, the workspace container has no resource constraints and can use as much of a given resource as the host’s kernel scheduler allows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    NuzeBot

    NuzeBot

    Finds interesting news headlines.

    This is a bot to finds the news you want to see. It can be made to find the news that interests you and reject everything else. View on one page the most interesting headlines from many websites.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3

    OGLDataScienceTool

    Opengl tool for data science visualization

    Data visualization tool written in LWJGL Compatible with libgdx and other opengl wrappers The project depends on apache poi, and apache commons, for office files support Planned features for next release: * reading json, and other nosql data structures * jdbc connection for creating dataframes * data heatmaps, and additional plots for questions, contact me kumar.santhi1982@hotmail.com more details: http://www.java-gaming.org/topics/ds/41920/view.html http://datascienceforindia.com/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Downloads: 0 This Week
    Last Update:
    See Project
  • Network Management Software and Tools for Businesses and Organizations | Auvik Networks Icon
    Network Management Software and Tools for Businesses and Organizations | Auvik Networks

    Mapping, inventory, config backup, and more.

    Reduce IT headaches and save time with a proven solution for automated network discovery, documentation, and performance monitoring. Choose Auvik because you'll see value in minutes, and stay with us to improve your IT for years to come.
    Learn More
  • 5

    Raku-DSL-Shared

    Raku package for DSL shared utilities and grammar roles.

    This repository provides a Raku package for shared utilities and (grammar) roles in the package context "DSL::". ("DSL" stands for "Domains Specific Language".) The initial versions of the code in this repository can be found in the GitHub repository [AAr1]. ## Utilities One of the reasons for making this package is to encapsulate and easily share utilities for making DSL translators. Here are "the first wave" utilities: Modify token patterns to include fuzzy matching Merge two or more roles into one ## Roles Another reason for making this package is to encapsulate and easily share grammar roles for making DSL translators. Here are "the first wave" roles: Error handling role Common English terms and phrases role used in workflows from Machine Learning, Data Science, or Scientific Computing Predicate specification role
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Recommenders

    Recommenders

    Best practices on recommendation systems

    The Recommenders repository provides examples and best practices for building recommendation systems, provided as Jupyter notebooks. The module reco_utils contains functions to simplify common tasks used when developing and evaluating recommender systems. Several utilities are provided in reco_utils to support common tasks such as loading datasets in the format expected by different algorithms, evaluating model outputs, and splitting training/test data. Implementations of several state-of-the-art algorithms are included for self-study and customization in your own applications. Please see the setup guide for more details on setting up your machine locally, on a data science virtual machine (DSVM) or on Azure Databricks. Independent or incubating algorithms and utilities are candidates for the contrib folder. This will house contributions which may not easily fit into the core repository or need time to refactor or mature the code and add necessary tests.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Rodeo

    Rodeo

    A data science IDE for Python

    A data science IDE for Python. RODEO, that is an open-source python IDE and has been brought up by the folks at yhat, is a development environment that is lightweight, intuitive and yet customizable to its very core and also contains all the features mentioned above that were searched for so long. It is just like your very own personal home base for exploration and interpretation of data that aims at Data Scientists and answers the main question, "Is there anything like RStudio for Python?" Rodeo makes it very easy for its users to explore what is created by them and also alongside allows the users to Inspect, interact, compare data frames, plots and even much more. It is an IDE that has been built especially for data science/Machine Learning in Python and you can also very simply think of it as a light weight alternative to the IPython Notebook.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    SageMaker Containers

    SageMaker Containers

    Create SageMaker-compatible Docker containers

    Amazon SageMaker is a fully managed service for data science and machine learning (ML) workflows. You can use Amazon SageMaker to simplify the process of building, training, and deploying ML models. To train a model, you can include your training script and dependencies in a Docker container that runs your training code. A container provides an effectively isolated environment, ensuring a consistent runtime and reliable training process. The SageMaker Training Toolkit can be easily added to any Docker container, making it compatible with SageMaker for training models. If you use a prebuilt SageMaker Docker image for training, this library may already be included. Very often, an entry point needs additional information from the container that is not available in hyperparameters. SageMaker Containers writes this information as environment variables that are available inside the script.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    SageMaker Inference Toolkit

    SageMaker Inference Toolkit

    Serve machine learning models within a Docker container

    Serve machine learning models within a Docker container using Amazon SageMaker. Amazon SageMaker is a fully managed service for data science and machine learning (ML) workflows. You can use Amazon SageMaker to simplify the process of building, training, and deploying ML models. Once you have a trained model, you can include it in a Docker container that runs your inference code. A container provides an effectively isolated environment, ensuring a consistent runtime regardless of where the container is deployed. Containerizing your model and code enables fast and reliable deployment of your model. The SageMaker Inference Toolkit implements a model serving stack and can be easily added to any Docker container, making it deployable to SageMaker. This library's serving stack is built on Multi Model Server, and it can serve your own models or those you trained on SageMaker using machine learning frameworks with native SageMaker support.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Inventory and Order Management Software for Multichannel Sellers Icon
    Inventory and Order Management Software for Multichannel Sellers

    Avoid stockouts, overselling, and losing control as your business grows.

    We are the most powerful inventory and order management platform for Amazon, Walmart, and multichannel product sellers. Centralize orders, product information, and fulfillment operations to run more efficiently, sell more products, and stay compliant with marketplace requirements so you can grow profitably.
    Learn More
  • 10
    Seldon Server

    Seldon Server

    Machine learning platform and recommendation engine on Kubernetes

    Seldon Server is a machine learning platform and recommendation engine built on Kubernetes. Seldon reduces time-to-value so models can get to work faster. Scale with confidence and minimize risk through interpretable results and transparent model performance. Seldon Core focuses purely on deploying a wide range of ML models on Kubernetes, allowing complex runtime serving graphs to be managed in production. Seldon Core is a progression of the goals of the Seldon-Server project but also a more restricted focus to solving the final step in a machine learning project which is serving models in production. Seldon Server is a machine learning platform that helps your data science team deploy models into production. It provides an open-source data science stack that runs within a Kubernetes Cluster. You can use Seldon to deploy machine learning and deep learning models into production on-premise or in the cloud (e.g. GCP, AWS, Azure).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Self-learning-Computer-Science

    Self-learning-Computer-Science

    Resources to learn computer science in your spare time

    Self-learning Computer Science is a curated, open-source guide repository designed to help learners independently study computer science topics using high-quality university-level resources. The author (an undergraduate CS student) assembled links to courses from institutions like MIT, UC Berkeley, Stanford, etc., covering mathematics, programming, data structures/algorithms, computer architecture, machine learning, software engineering and more. It’s aimed at learners who find traditional course structures restrictive and want a flexible, self-paced path through CS, with a focus on building depth and breadth rather than shortcut exam skills. The repository provides a roadmap, references, teaching materials, and sometimes the author’s own project examples, offering both guidance and community support. Because the CS field is broad, the structure helps learners allocate study time, avoid duplication, and benefit from “best in class” resources instead of randomly browsing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Spark Notebook

    Spark Notebook

    Interactive and Reactive Data Science using Scala and Spark

    Spark Notebook is an interactive web-based computational notebook designed to make working with Apache Spark more productive, exploratory, and expressive. It allows developers, data scientists, and analysts to write, run, and visualize Spark code in cells that support multiple languages such as Scala, Python, and SQL, all within the same notebook. Users can interleave runnable code, rich text markup, visualizations, equations, and results, enabling reproducible research and exploratory data analysis workflows. Because it runs on top of Spark’s distributed engine, it can scale from running locally on a laptop to executing on clusters with large datasets without changing user workflow. The UI is notebook-style with support for incremental execution, error inspection, and stateful session continuity, making it easy to iterate on data transformations and model training tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Synapse Machine Learning

    Synapse Machine Learning

    Simple and distributed Machine Learning

    SynapseML (previously MMLSpark) is an open source library to simplify the creation of scalable machine learning pipelines. SynapseML builds on Apache Spark and SparkML to enable new kinds of machine learning, analytics, and model deployment workflows. SynapseML adds many deep learning and data science tools to the Spark ecosystem, including seamless integration of Spark Machine Learning pipelines with the Open Neural Network Exchange (ONNX), LightGBM, The Cognitive Services, Vowpal Wabbit, and OpenCV. These tools enable powerful and highly-scalable predictive and analytical models for a variety of data sources. SynapseML also brings new networking capabilities to the Spark Ecosystem. With the HTTP on Spark project, users can embed any web service into their SparkML models. For production-grade deployment, the Spark Serving project enables high throughput, sub-millisecond latency web services, backed by your Spark cluster.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    TensorWatch

    TensorWatch

    Debugging, monitoring and visualization for Python Machine Learning

    TensorWatch is an open source debugging and visualization platform created by Microsoft Research to support machine learning, deep learning, and reinforcement learning workflows. It enables developers to observe training behavior in real time through interactive visualizations, primarily within Jupyter Notebook environments. The tool treats most data interactions as streams, allowing flexible routing, storage, and visualization of metrics generated during model training. A distinctive capability is its “lazy logging” mode, which lets users query live training processes without pre-instrumenting all metrics ahead of time. TensorWatch supports multiple chart types and can be extended with custom visualizers and dashboards, making it highly adaptable for research workflows. Overall, the project acts as a powerful observability layer for ML experimentation, helping practitioners diagnose model behavior and compare runs more efficiently.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15

    Ubuntu -16.04-DataScience-stack

    To provide a customized environment to practice data science

    Although its a relatively easy task to setup, a customized environment to practice data science with the python tool stack is less common, including this site, Vagrant boxes and osboxes.org. Hence this project is kicked out as of early 2019.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    alltools

    The Data Science Toolbox

    Downloads: 0 This Week
    Last Update:
    See Project
  • 17

    slycat

    Web-based data science analysis and visualization platform.

    This is Slycat - a web-based data science analysis and visualization platform, created at Sandia National Laboratories. The goal of the Slycat project is to develop processes, tools and techniques to support data science, particularly analysis of large, high-dimensional data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    targets

    targets

    Function-oriented Make-like declarative workflows for R

    The targets package is a pipeline / workflow management tool in R, designed to coordinate multi‐step computational workflows in data science / statistics. It tracks dependencies between “targets” (computational steps), skips steps whose upstream data or code hasn’t changed, supports parallel computation, branching (dynamic generation of sub‐targets), file format abstractions, and encourages reproducible and efficient analyses. It’s something like GNU Make for R, but more integrated. Skipping computation for up-to-date targets so that unchanged parts of the workflow are not recomputed. Targets can represent files or R objects, and tracking file changes etc is incorporated.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB