Showing 2 open source projects for "how to use hakuneko s"

View related business solutions
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • Inventory and Order Management Software for Multichannel Sellers Icon
    Inventory and Order Management Software for Multichannel Sellers

    Avoid stockouts, overselling, and losing control as your business grows.

    We are the most powerful inventory and order management platform for Amazon, Walmart, and multichannel product sellers. Centralize orders, product information, and fulfillment operations to run more efficiently, sell more products, and stay compliant with marketplace requirements so you can grow profitably.
    Learn More
  • 1
    tapir

    tapir

    Declarative, type-safe web endpoints library

    ...Each endpoint can contain a number of input and output parameters. Compile-time guarantees, develop-time completions, read-time information. Separate the shape of the endpoint (the "what"), from the server logic (the "how"). Generate documentation from endpoint descriptions. Leverage the metadata to report rich metrics and tracing information. Re-use common endpoint definitions, as well as individual inputs/outputs. Library, not a framework, integrates with your stack. Is your company already using tapir? We're continually expanding the "adopters" section in the documentation; the more the merrier! ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    apache spark data pipeline osDQ

    apache spark data pipeline osDQ

    osDQ dedicated to create apache spark based data pipeline using JSON

    This is an offshoot project of open source data quality (osDQ) project https://sourceforge.net/projects/dataquality/ This sub project will create apache spark based data pipeline where JSON based metadata (file) will be used to run data processing , data pipeline , data quality and data preparation and data modeling features for big data. This uses java API of apache spark. It can run in local mode also. Get json example at https://github.com/arrahtech/osdq-spark How to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB