-
### Your current environment
The output of `python collect_env.py`
```text
PyTorch version: 2.4.0+cu121
OS: Ubuntu 22.04.3 LTS (x86_64)
Python version: 3.10.12 (main, Nov 20 2023, 15:14:05…
-
I am getting the following error when I try running `report_numerical_issues()` from the DiagnosticToolbox. I should note that this was working fine until I updated my WaterTAP environment--just haven…
-
I have a question: how to parallel your model using BERT?
-
In the README the following warning can be read:
"Note that the Anthropic API, llama-server (and ollama) currently does not support sampling multiple responses from a model, which limits the availa…
-
### Prerequisites
- [X] I have read the [ServerlessLLM documentation](https://serverlessllm.github.io/).
- [X] I have searched the [Issue Tracker](https://github.com/ServerlessLLM/ServerlessLLM/issue…
-
### Description
As of astropy 5.3 I get the following error when building the `dkist` documentation with parallel jobs:
```
Process ForkProcess-19:
Traceback (most recent call last)…
-
### Description
```python
fitter = TRFLSQFitter(calc_uncertainties=True)
spice_model_fit = parallel_fit_dask(
data=spice,
model=average_fit,
fitter=fitter,
fitting_axes=0,
…
-
Hie,
Thanks for opensourcing this amazing work. Is there any parameter to parallize the model to run on smaller gpus. I was not able to find one in config. As suggested in readme "we should turn on m…
-
### My Python Script
```python
import os
os.environ["CUDA_VISIBLE_DEVICES"] = "2,3"
import time
import torch
from vllm import LLM, SamplingParams
import gc
from vllm.model_executor.parallel_…
-
I deployed Qwen-2VL-72B using Swift, but during multi-image content inference, the generated results consistently terminate early. Could you advise on how to resolve this?
The startup script is as …