-
Dear Shankar1729,
Thank you for creating such an intuitive and powerful program.
I have run into an issue that I have been trying to debug for a couple months to no avail. I hope that you may hav…
-
try two methods to create container (ghcr.io/huggingface/text-generation-inference:sha-7de8a37) but no luck:
```
docker run --gpus all --shm-size 1g -p 8080:80 ghcr.io/huggingface/text-generation…
-
**Summary**
Got error about Cuda driver
mpirun --allow-run-as-root -n 32 lmp -sf gpu -pk gpu 2 -restart2data lmp.restart remap lmp_final.data
![image](https://github.com/lammps/lammps/assets/1…
-
Frameworks
[VLDB '20] PyTorch Distributed: Experiences on Accelerating Data Parallel Training
[NeurIPS '19] PyTorch: An Imperative Style, High-Performance Deep Learning Library
[OSDI '18] Ray: A Di…
-
I did some tests on faster GPUs (Nvidia 3090) with current pipelines, and even if the feature extraction is in the network the data loading is blocking the training (GPU utilization in nvidia-smi only…
-
# Replace Behat with Cypress
> **ESTIMATE** TBD
Behat is a behavioral testing language that uses Gherkin for test features, and PHP to handle step definitions. Within the BLT stack, Behat is used …
-
Dear Marius,
We, in Andreas Tolias Lab at Baylor College of Medicine in Houston Texas, are planning to do close to 2 hours worth of recordings using neuropixels in mouse. Based on some calculations…
-
Can you provide an example of how to use `accelerate` with the [Hugging Face trainer](https://huggingface.co/transformers/master/main_classes/trainer.html#id1)?
-
Thanks for creating these awesome kernels! I am trying to get flashinfer kernels to work with cuda graphs. But it appears that several parallelism decisions (block size, num_q_tiles, etc.) are made on…
-
As I work through [this](https://github.com/OpenNMT/OpenNMT-py/tree/master/docs/source/examples/wmt17) example, I soon discovered the following error message:
`This fp16_optimizer is designed to o…