-
# Trending repositories for C#
1. [**dotnet-architecture / eShopOnWeb**](https://github.com/dotnet-architecture/eShopOnWeb)
__Sample ASP.NET Core 6.0 reference application, powere…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
问了几个问题,为什么生成的回答不完整呢
### Expected Behavior
_No response_
### Steps To Reproduce
这样不完…
-
### Is your feature request related to a problem? Please describe.
目前能看到各种微调的方式比如p-tuning和lora,请问如果我想像pretrain那样,直接mask文章的片段去训练模型,需要怎么做呢?
### Solutions
能否直接给出训练的demo?包括需要怎么准备训练数据,如何mask之类,十分感谢
###…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
您好,我这边在尝试基于 ChatGLM-6B在领域数据上先做自监督增量预训练,然后再做指示微调。有几个问题想请教下您,望赐教:
1.您认为此方案可行性如何,基于ChatGLM-6B再做…
-
实现了 ChatGLM2-6B 的lora微调,可以用来做领域微调,它的SFT微调方法跟chatglm基本一致,需要改下special tokens, lm_head 和enable_input_require_grads 就可以适配(下面代码改好了的)。
支持THUDM/chatglm2-6b微调项目地址:https://github.com/shibing624/MedicalGPT
…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
我用peft的lora在 chatglm2 训练了一个Lora(参考了 @shibing624 https://github.com/shibing624/MedicalGPT 里面的方…
-
While finetuning RWKV, I use this script(using demo dataset by `make_data.py` and put `demo.bin` and `demo.idx` in `./data`):
```
#!/bin/bash
BASE_NAME="model/demo"
N_LAYER="12"
N_EMBD="768"
M…
-
以ChatGLM为基座,增加自己的语料库进行预训练,不是微调,这个该怎么做?
-
### Describe the Question
Please provide a clear and concise description of what the question is.
大佬您好,我现在在使用qwen2-7b-instruct在做领域微调,打算采用pt+sft的路线,但是我在做完pt后,使用ceval数据集做测试的时候,在电力领域为什么得分反而比基座模型的得分低。…
-
### Describe the Question
Please provide a clear and concise description of what the question is.
徐老师,运行您给的notebook报错
![image](https://github.com/shibing624/MedicalGPT/assets/72805517/560cea89-3abc…