-
I meets the errors when funtune using loar.
ValueError: Target module LlamaDecoderLayer(
(self_attn): LlamaAttention(
(q_proj): Linear(in_features=2048, out_features=2048, bias=False)
(k…
-
Post questions here for this week's fundamental readings:
J. Evans and B. Desikan. 2022. “Deep Learning?” and “Deep Neural network models of text”, Thinking with Deep Learning, chapter 1, 9
Ash…
lkcao updated
8 months ago
-
### Feature request / 功能建议
This feature request proposes adding support for Meta's newly released Llama 3.2 models to lmdeploy. Llama 3.2 introduces exciting capabilities, including vision LLMs (11…
-
Context:
@snat-s has done great work w/ the analysis of various data that may be relevant to Neko
We should now, w/ the input of the team, finalize our proposed V0 dataset, justify its
Output: doc…
-
### System Info
- `transformers` version: 4.45.2
- Platform: Linux-5.15.0-120-generic-x86_64-with-glibc2.35
- Python version: 3.10.15
- Huggingface_hub version: 0.25.2
- Safetensors version: 0.4.…
-
## Environment
FastDeploy version: latest code in develop branch
OS Platform: Linux (Linux 5.10.110-rockchip-rk3588 #23.02.2 SMP Fri Feb 17 23:59:20 UTC 2023)
Hardware: e.g. Orange Pi 5 Rockchip …
-
For ollama and openai compatible endpoints, use the syntax to force JSON output.
For LMStudio served models specifically, I can force a stronger JSON schema with defined validation logic.
-
-
Project Lead: @tedarbyshire
Mentor: @joelostblom
Welcome to OL6, Cohort B! This issue will be used to track your project and progress during the program. Please use this checklist over the next fe…
-
#### Environment details
- OS type and version: Mac OS, Python
#### Steps to reproduce
1. Include a Tool in the GenerativeModel.generate_content call. Don't specify any System Instruction…