-
Any plans to support conversion of ```StableLmForCausalLM" models? I've noticed that they're very good; for example the new Zephyr model here:
https://huggingface.co/stabilityai/stablelm-zephyr-3b…
-
### Enhancement Request - Support for Additional LLM Types
#### Description:
After reviewing the [Mediapipe documentation](https://developers.google.com/mediapipe/solutions/genai/llm_inference) an…
-
### Please check that this issue hasn't been reported before.
- [X] I searched previous [Bug Reports](https://github.com/OpenAccess-AI-Collective/axolotl/labels/bug) didn't find any similar reports.
…
-
### Feature request
Generalize the functionality in [processing_llava.py](https://github.com/huggingface/transformers/blob/main/src/transformers/models/llava/processing_llava.py) to include other t…
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### W…
-
Hi,
I am running llama-cpp-python on surface book 2 having i7 with nvidea geforce gtx 1060.
I installed vc++, cuda drivers 12.4
Running on Python 3.11.3
Compiled llama using below command on Min…
-
I found that the scripts in GEMMA do not support GEMMA2. Is there any plan to add support for GEMMA2?
-
1) Load the Gemma2 2B model with Unsloth - OK
2) Perform fine tuning - OK
3) Test the resulting model - OK, responses indicate fine tuning is successful
4) Save 16 bit `model.save_pretrained_merged…
-
Doesn't seem to work for these new models currently. I'm on a Mac and get the following error for these models:
ValueError: Can't infer missing attention mask on `mps` device. Please provide an `atte…
-
Ref: https://github.com/ggerganov/llama.cpp/pull/8687#issuecomment-2252155218
(cc @ggerganov)
TODO:
- Train some adapters based on stories15M and [stories15M_MOE](https://huggingface.co/ngxson/…