predibase / lorax

Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
https://loraexchange.ai
Apache License 2.0
1.86k stars 126 forks source link

Project Roadmap #57

Open tgaddair opened 7 months ago

tgaddair commented 7 months ago

WIP project roadmap for LoRAX. We'll continue to update this over time.

v0.10

v0.11

Previous Releases

v0.9

Backlog

Models

Adapters

Throughput / Latency

Quantization

Usability

RileyCodes commented 7 months ago

is AWQ supported?

tgaddair commented 7 months ago

Hey @RileyCodes, not yet, will add that to the roadmap!

abhibst commented 7 months ago

does we have tested bitsandbytes Quantization ?

tgaddair commented 7 months ago

Hey @abhibst, I've done some basic sanity checks on it, but haven't tested it very thoroughly. Please feel free to report any issues you encounter and I'll take a look!

abhibst commented 7 months ago

Sure Thanks for confirming

sansavision commented 7 months ago

How would you go about adding this in Stable Diffusion? I am really interested in experimenting with that.

tgaddair commented 7 months ago

Hey @sansavision, at a high level it would look a lot like the LoRA pipeline used in Diffusers: https://github.com/huggingface/api-inference-community/blob/main/docker_images/diffusers/app/pipelines/text_to_image.py#L25

A v0 shouldn't be too bad, we would basically just run a single forward pass to generate the image and perform postprocessing (as part of the existing Prefill step) and short-circuit the Decode step.

flozi00 commented 7 months ago

If no one has started I will start working on awq tomorrow

tgaddair commented 7 months ago

Nice! Thanks @flozi00, that would be awesome!

SamGalanakis commented 6 months ago

Any plans to support vision transformers from huggingface / timm? A lot of potential use cases there for deploying many classifiers. If not what would that entail? Would be open to contributing if possible.

tgaddair commented 6 months ago

Hey @SamGalanakis, great suggestion! The plan at the moment is to start by supporting text classifiers. Once that framework is in place for that, it should be hopefully relatively straightforward to support image classifiers as well. Happy to start a thread on Discord to discuss!

flozi00 commented 6 months ago

Whisper would be also very cool 😄

SamGalanakis commented 6 months ago

@tgaddair Ok clear, joined the discord will look out for it!

Hap-Zhang commented 6 months ago

Hi, @tgaddair , could I know how long it will take to support the stable diffusion model?

tgaddair commented 6 months ago

Hey @Hap-Zhang, the plan at the moment is to add it after we add support for embedding generation and text classification. Both of those are planned for January 2024, so in the next month.

Hap-Zhang commented 6 months ago

@tgaddair Okay, got it. Thank you very much for your efforts. Stay tuned for it.

AdithyanI commented 5 months ago

If we could have OpenAI compatible endpoints that would be great too. So we can use this as drop in replacement for OpenAI models :)

tgaddair commented 5 months ago

Hey @AdithyanI, yes, this should be coming this week or next! See #145 to follow progress.

AdithyanI commented 5 months ago

@tgaddair oh wow that would be awesome! Thank you so much for the work here. If you need someone to test it out; let me know. Happy to test it out.

Is the discord still open for others to join :) ? I followed the link of the repo, and it says it is expired.

tgaddair commented 5 months ago

@AdithyanI this should be landing some time today :)

170

tgaddair commented 5 months ago

Hey @AdithyanI, the Discord should be available. Are you using this link?

https://discord.gg/CBgdrGnZjy

AdithyanI commented 5 months ago

@tgaddair I asked for outlines repo authors to add support to this : https://github.com/outlines-dev/outlines/discussions/523 Then it would be great to have text guided generation :)

image

I don't know how hard is it to integrate that here. Do you folks know if this is something that can be supported by LORAX?

tgaddair commented 5 months ago

Thanks for starting the Outlines thread @AdithyanI! Looks like the maintainer created an issue #176. Excited to explore this integration!

K-Mistele commented 4 months ago

Would it be possible to add in context length-scaling methods like Self-Extend , Rope scaling, and/or yarn scaling? I know that llama.cpp has a good implementation of these in their server, and self-extend in particular is much more stable than rope or yarn. Having long context or doing context enhancement is super important for RAG applications.

thincal commented 4 months ago

About the supported models, could you consider the ChatGLM3 ? @tgaddair

thincal commented 3 months ago
  • LongLoRA

It seems that LongLoRA proposed shifted short attention is compatible with Flash-Attention, and not required during inference (ref: https://huggingface.co/Yukang/Llama-2-13b-longlora-8k#highlights), if that is true, could you share what's the planed support in LoRAX inference side? thanks @tgaddair

remiconnesson commented 3 months ago

Do you plan on supporting AQLM to setve LoRa of Mixtral Instruct with Lorax?

tgaddair commented 3 months ago

Hey @thincal, the last thing we need to support LongLoRA, if I remember correctly, is #231 which @geoffreyangus is planning to pick up next week.

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

amir-in-a-cynch commented 3 months ago

Are T5 based models on the Roadmap?

remiconnesson commented 3 months ago

@tgaddair

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

Hello :) How far do you think we are for this PR to be merged? :)

tgaddair commented 3 months ago

Hey @remiconnesson, will probably be the next thing I take a look at after wrapping up speculative decoding this week.

@amir-in-a-cynch we can definitely add T5 to the roadmap!

tomrance commented 2 months ago

Hello, will you integrate / merge / migrate to the latest hugging face text-generation-inference as it is back now with Apache 2.0 license?