-
Hey there,
I need to evaluate the (max) memory usage and annotation speed, for the following annotators:
- Tokenization into words
- Tokenization into sentences
- Part-of-speech tagging
- L…
-
While watching a few thousand files copy from a memory card with -g I was wishing there was an overall average speed display on the top progress bar, to go along with the per-file speed on the lower o…
-
# ❓ Questions and Help
I am new to xformers, and I want to speed my Transformer models w/ it. But I found that `xformers` is no speed up compared w/ `scaled_dot_product_attention` from PyTorch. Here …
-
I am running the `inspect_packages` pipeline on a very large codebase that is an npm package. The pipeline takes a very long time at the `scan_for_application_packages` step on the package assembly po…
-
Hi, I recently converted my robot from urdf to instanceable usd. I double checked if the actual usd is instanceable and printed out the prim.IsInstanceable(). However, I tested the speed in source/sta…
-
Currently, in lmcache_vllm, we do memory copy and `tensor.contiguous`
Should profile the overhead and estimate what could be the gain.
-
### System Info
Dear authors,
I have a question regarding the training time utilizing the peft package. I tried using LoRA with a swin transformer to reduce the parameter size.
```
model = Swi…
-
No package located for Ubuntu 24
`ubuntu@rpi5-orso-1:~$ sudo apt install hailo-all
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
E: Unable to lo…
-
The `container_*` Prometheus metrics are missing on certain systems, can someone help?
For instance, I see: `container_cpu_usage_usec_microseconds` and `container_memory_usage_bytes` on an Ubuntu 22.…
-
**Hello! I use auto-gptq to quantized `llama-2-7b-instruct` model to `llama-2-7b-instruct-4bit-128g`. And i try to compare the speed between them. But the result is very strange. The storage of the qu…