TBD-Labs-AI / latched

AI on the Go
Apache License 2.0
13 stars 6 forks source link

Latched

PRs Welcome License Website

ML models latches onto devices

Latched provides easy-to-use pipelines to perform ML models on various devices such as mobile, Nvidia jetson, Intel CPUs, and accelerators. Latched covers both converting models and deploying them(Latched Model Manager, Latched Devices SDKs).

🤖 Supported ML Tasks

📚 Text:

🏞️ VIsion:

🗣️ Audio:

Supported Frameworks:

🧩 Latched Components

Latched: Latched python library provides hardware-aware optimization. With this library, you can export your ML model into hardware-optimized forms.

Latched Model Manager: Latched model manager provides a RESTful API to register and run ML models on various devices.

Latched Devices SDKs: Latched devices SDKs provide libraries to run ML models on various devices.

🚀 Getting Started

Installation

  1. Clone the repository
    git clone https://github.com/TBD-Labs-AI/latched.git
    cd latched
  2. Make the virtual environment with Python 3.11.9 and activate it.
    conda create -n latched python=3.11.9
    conda activate latched
  3. Install the dependencies with Poetry
    pip install poetry
    poetry install
  4. Launch the test script (onnx export)
    python examples/llama-3.1-8B-Instruct-to-onnx/llama_onnx_example.py

How to use Latched

  1. Export HuggingFace Models to the ONNX format
  2. Export HuggingFace Models to the OpenVINO format

📚 Model Hub

coming soon

Contributing

Do you believe the future of AI is on edge computing? Do you want to make it happen? Join Latched as a contributor! If you want to contribute to Latched, please read the CONTRIBUTING.md file.

📅 Milestones

SEP 2024

This repository uses the following third-party libraries: