-
**Dear Gents,**
**I am trying to fine-tune the VGG16 model in order to classify 5 classes, inspired by the link :** https://gist.github.com/fchollet/7eb39b44eb9e16e59632d25fb3119975.
**When i u…
-
### 请提出你的问题
自己尝试llama做predict,遇到问题
python -u -m paddle.distributed.launch \
--gpus "6,7" \
--log_dir "output/$task_name""_log" \
run_pretrain.py \
--model_type "llama" \
--…
-
### 🐛 Describe the bug
Hello,
I tried using the beta forward-mode automatic differentiation, but I ran into an issue when trying to compile my forward pass. I wonder if it was an error on my part or…
-
**Description:** I am experiencing issues using my GPU (Quadro K2200) with the latest software. Below is the log output when I try to load a model.
**Steps Taken:**
1. Initially, I was using the…
-
### 🐛 Describe the bug
Using nested tensors generated with `torch.narrow` as inputs to `torch.nn.functional.scaled_dot_product_attention` works fine in the forward pass of the model. However, both …
-
We are trying to use CellProfiler to batch analyze several hundred folders worth of images. Each folder has ~2500 images; each image is 2.3 MB. Currently we’re dragging and dropping a batch of folders…
-
### System Info
- Platform: Linux-5.15.0-86-generic-x86_64-with-glibc2.35
- Python version: 3.10.14
- PyTorch version: 2.4.1
- CUDA device: NVIDIA A100-SXM4-80GB
- Transformers version: 4.45.0.…
-
### 🐛 Describe the bug
Functions [`torch.nn.functional.relu`](https://github.com/pytorch/pytorch/blob/main/torch/nn/functional.py#L1693) and [`torch.nn.functional.multi_head_attention_forward`](https…
-
**Describe the bug**
I use llama-2 7b, and when I start stage 2 in EE-Tuning, the bug occurs.
**To Reproduce**
here is `llama2_7B_1_exit_mlp_pt.sh` I modified:
``` bash
#!/bin/bash
PROJECT…
-
### 🐛 Describe the bug
Enabling both these options causes an error.
Code:
```
import torch
batch_size = 32
seq_length = 50
hidden_size = 768
def test_fn():
inp = torch.randn(batch_size,…