predibase / lorax

Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
https://loraexchange.ai
Apache License 2.0
2.22k stars 145 forks source link

Project Roadmap #57

Open tgaddair opened 1 year ago

tgaddair commented 1 year ago

WIP project roadmap for LoRAX. We'll continue to update this over time.

v0.10

v0.11

Previous Releases

v0.9

Backlog

Models

Adapters

Throughput / Latency

Quantization

Usability

RileyCodes commented 1 year ago

is AWQ supported?

tgaddair commented 1 year ago

Hey @RileyCodes, not yet, will add that to the roadmap!

abhibst commented 1 year ago

does we have tested bitsandbytes Quantization ?

tgaddair commented 1 year ago

Hey @abhibst, I've done some basic sanity checks on it, but haven't tested it very thoroughly. Please feel free to report any issues you encounter and I'll take a look!

abhibst commented 1 year ago

Sure Thanks for confirming

sansavision commented 12 months ago

How would you go about adding this in Stable Diffusion? I am really interested in experimenting with that.

tgaddair commented 12 months ago

Hey @sansavision, at a high level it would look a lot like the LoRA pipeline used in Diffusers: https://github.com/huggingface/api-inference-community/blob/main/docker_images/diffusers/app/pipelines/text_to_image.py#L25

A v0 shouldn't be too bad, we would basically just run a single forward pass to generate the image and perform postprocessing (as part of the existing Prefill step) and short-circuit the Decode step.

flozi00 commented 12 months ago

If no one has started I will start working on awq tomorrow

tgaddair commented 12 months ago

Nice! Thanks @flozi00, that would be awesome!

SamGalanakis commented 11 months ago

Any plans to support vision transformers from huggingface / timm? A lot of potential use cases there for deploying many classifiers. If not what would that entail? Would be open to contributing if possible.

tgaddair commented 11 months ago

Hey @SamGalanakis, great suggestion! The plan at the moment is to start by supporting text classifiers. Once that framework is in place for that, it should be hopefully relatively straightforward to support image classifiers as well. Happy to start a thread on Discord to discuss!

flozi00 commented 11 months ago

Whisper would be also very cool 😄

SamGalanakis commented 11 months ago

@tgaddair Ok clear, joined the discord will look out for it!

Hap-Zhang commented 11 months ago

Hi, @tgaddair , could I know how long it will take to support the stable diffusion model?

tgaddair commented 11 months ago

Hey @Hap-Zhang, the plan at the moment is to add it after we add support for embedding generation and text classification. Both of those are planned for January 2024, so in the next month.

Hap-Zhang commented 11 months ago

@tgaddair Okay, got it. Thank you very much for your efforts. Stay tuned for it.

AdithyanI commented 10 months ago

If we could have OpenAI compatible endpoints that would be great too. So we can use this as drop in replacement for OpenAI models :)

tgaddair commented 10 months ago

Hey @AdithyanI, yes, this should be coming this week or next! See #145 to follow progress.

AdithyanI commented 10 months ago

@tgaddair oh wow that would be awesome! Thank you so much for the work here. If you need someone to test it out; let me know. Happy to test it out.

Is the discord still open for others to join :) ? I followed the link of the repo, and it says it is expired.

tgaddair commented 10 months ago

@AdithyanI this should be landing some time today :)

170

tgaddair commented 10 months ago

Hey @AdithyanI, the Discord should be available. Are you using this link?

https://discord.gg/CBgdrGnZjy

AdithyanI commented 10 months ago

@tgaddair I asked for outlines repo authors to add support to this : https://github.com/outlines-dev/outlines/discussions/523 Then it would be great to have text guided generation :)

image

I don't know how hard is it to integrate that here. Do you folks know if this is something that can be supported by LORAX?

tgaddair commented 10 months ago

Thanks for starting the Outlines thread @AdithyanI! Looks like the maintainer created an issue #176. Excited to explore this integration!

K-Mistele commented 9 months ago

Would it be possible to add in context length-scaling methods like Self-Extend , Rope scaling, and/or yarn scaling? I know that llama.cpp has a good implementation of these in their server, and self-extend in particular is much more stable than rope or yarn. Having long context or doing context enhancement is super important for RAG applications.

thincal commented 9 months ago

About the supported models, could you consider the ChatGLM3 ? @tgaddair

thincal commented 8 months ago
  • LongLoRA

It seems that LongLoRA proposed shifted short attention is compatible with Flash-Attention, and not required during inference (ref: https://huggingface.co/Yukang/Llama-2-13b-longlora-8k#highlights), if that is true, could you share what's the planed support in LoRAX inference side? thanks @tgaddair

remiconnesson commented 8 months ago

Do you plan on supporting AQLM to setve LoRa of Mixtral Instruct with Lorax?

tgaddair commented 8 months ago

Hey @thincal, the last thing we need to support LongLoRA, if I remember correctly, is #231 which @geoffreyangus is planning to pick up next week.

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

amir-in-a-cynch commented 8 months ago

Are T5 based models on the Roadmap?

remiconnesson commented 8 months ago

@tgaddair

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

Hello :) How far do you think we are for this PR to be merged? :)

tgaddair commented 7 months ago

Hey @remiconnesson, will probably be the next thing I take a look at after wrapping up speculative decoding this week.

@amir-in-a-cynch we can definitely add T5 to the roadmap!

tomrance commented 7 months ago

Hello, will you integrate / merge / migrate to the latest hugging face text-generation-inference as it is back now with Apache 2.0 license?

bdalal commented 3 months ago

Is there an expected release date for v0.11?