-
### Feature Description
Integrate AI21Labs' Jamba 1.5 models into the Vercel AI SDK, leveraging their 256K token context window and hybrid architecture for enhanced long-context understanding and eff…
-
### Checklist
- [x] 1. If the issue you raised is not a feature but a question, please raise a discussion at https://github.com/sgl-project/sglang/discussions/new/choose Otherwise, it will be closed.…
-
[Jamba](https://huggingface.co/ai21labs/Jamba-v0.1) is a very interesting new model and I’d love to add support for galore for finetuning it. It’s an MoE+Transformer+Mamba hybrid so I’m not sure how t…
-
PyTorch is dead. Long live JAX.
https://neel04.github.io/my-website/blog/pytorch_rant/
LLM Compressor
https://github.com/vllm-project/llm-compressor
https://neuralmagic.com/blog/llm-compressor-i…
-
### Model ID
ai21labs/Jamba-v0.1
### Model type
Decoder model (e.g., GPT)
### Model languages
- [x] Danish
- [x] Swedish
- [x] Norwegian (Bokmål or Nynorsk)
- [x] Icelandic
- [x] Faroese
- [x] Ge…
-
- Advanced type of Language Model using Deep learning techniques using heavy text data.
- Capable of generating human like text. QnA, Text2Text
- Concepts like n-gram to Neural Networks are used. …
-
Can I use this solution for inference https://huggingface.co/ai21labs/Jamba-v0.1/discussions with offloading mamba moe layers?
Jambo it SOTA open source long context model and its support would be …
-
### What model would you like?
https://www.maginative.com/article/ai21-labs-unveils-jamba-the-first-production-grade-mamba-based-ai-model/
https://huggingface.co/ai21labs/Jamba-v0.1?ref=maginati…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
Hey! I'm really impressed by your work on VLM. It has a huge impact for the MLLM area.
I wonder about the LLM in your VLM. Did you train it from 0? Or use a pre - trained model like "ai21labs/Jamba -…