-
Now that vulkan gpu acceleration has been merged into llama.cpp is it possible to implement it here? It may give a good performance boost to the local backend.
-
Why do I am getting "Could not reach the server"?
-
When I run the fine-tuning code on my dataset I get this error:
TypeError: `model` must be a `LightningModule` or `torch._dynamo.OptimizedModule`, got `LagLlamaLightningModule`
But the zero-shot…
-
### What is the issue?
Since I installed ollama (v0.1.30) on Windows 11 Pro, I run into a peculiar issue. When I launch ollama from the installed shortcut, which launches "ollama app.exe", or when …
-
is there a way to run with AMD GPU ?
-
I'm currently working on the tokenizer, we need a new one.
The llama tokenizer is not suitable, it has problems forming larger tokens and favors smaller ones and it does not adhere to the merge pri…
-
I need a way to dynamically set the rate limit based on the user type.
For example, I want to limit users without access token & have unlimited access to users with the access token.
What I am …
-
I noticed that the dimensions of the output of the first term in the list of:
`forecast_it, ts_it = make_evaluation_predictions(
dataset=backtest_dataset,
predictor=predictor,
)`
is a…
-
File [~/miniconda3/envs/env/lib/python3.11/site-packages/pytorch_lightning/trainer/trainer.py:580](https://file+.vscode-resource.vscode-cdn.net/home/seyed/PycharmProjects/step/lag-llama/~/miniconda3/e…
-
Example Using Automatic Prompt Generation
Default settings
Prompt executed in 498.52 seconds
The graphics card uses 3090.
CPU usageIntel(R) Xeon(R) Platinum 8362 CPU