-
In reflection-classification/models/README.md, the link in the gdown snipet should be updated.
Original line:
`gdown https://drive.google.com/uc?id=1Sv0OpLyi13HA5miRxbi5HNGmziJ0yQwt -O xlm-roberta…
-
Thank you for providing such a great resources!
I was wondering if it is possible to use LoRA (Low-Rank Adaptation) to reduce the resource requirements and potentially fine-tune even larger models.
…
-
[Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback](https://arxiv.org/abs/2302.12813)
[Active Prompting with Chain-of-Thought for Large La…
-
Have you ever tried a model with a larger number of parameters,for example, 30M,100M?I don't know how good the performance is compared to models like convnext.
-
# Issues Discription #
I tried to use the following command to run all model level tests:
```
python run.py -j 16 --report --cachedir cached -v --testsfile models.txt \
--torchmlirbuild /torch…
-
Hey!
Great job on Arena, I think in the era of saturated benchmarks, having an actual large-number vibes-based evaluation is very important.
I was wondering, would you entertain adding models th…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
### What is the issue?
I have noticed that when GPU VRAM gets near-full, but ollama has decided to load 2 models into VRAM, incoming requests to one model simply stall until the other model pops out …
-
For some of our work we would like to estimate models with large numbers of fixed effects. Currently `eventstudyr` uses `plm`under the hood, but it would be useful to be able to use `fixest` too.
-
So the Sharp X68000 has two different variations of the hard disk controller bus: the Super/XVI/Compact/X68030 use the SCSI bus. However the older X68000 revisions (CZ-600C/Pro/Expert/Ace) all use the…