-
I am trying to fine tune Mistral 7B model and want to use PEFT setting. but getting following error :
File [~/.conda/envs/env/lib/python3.10/site-packages/dspy/teleprompt/finetune.py:166](http://lo…
-
I'm possibly doing something a bit naive, but i just wondered if there was interpretation functions available for spline based models like you get with mgcv in R.
```python
import pandas as pd
i…
-
In Hugging Face "eager" Mistral implementation, a sliding window of size 2048 will mask 2049 tokens. This is also true for flash attention. In the current vLLM implementation a window of 2048 will mas…
caiom updated
1 month ago
-
I am trying to run the bcsd example and the following code
>>>out = bcsd_temp.predict(X_temp) + X_temp
gives the error as below
AttributeErrorTraceback (most recent call last)
in
…
-
Hi, I'm Chang-Bin from Seoul National University.
Recently, I made the neural De-limiter (https://github.com/jeonchangbin49/De-limiter), which estimate the uncompressed music signals from heavily c…
-
so far we don't have any models that allow for endogenous regressors (outside of sandbox)
base GMM classes need to be somewhere central to avoid circular imports.
Models that handle endogenous r…
-
## ❓ Questions and Help
1) The baseline in your paper also has two types: baseline with SUM, baseline with GATE. Can you provide the training and test command for these two baselines (baseline with…
-
It seems to me that BSRNN with `causal=True` is not causal. Forward passing a tensor with NaNs in the last time frame results in a tensor with NaNs only.
```python
import torch
from espnet2.enh…
-
Falied to finetune finetune_guanaco_7b:
```
File "/home/ubuntu/anaconda3/envs/qlora/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 465, in _prepare_decoder_atten …
-
### System Info
lorax latest docker, 2 A100, unbuntu 22.04
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [ ] An officially supported command
- [ ] My own modifications
### Repr…