-
The interface for Slicer N4ITKBiasFieldCorrection does not work with Slicer 4.10. The input/output files are now positional arguments and not flags. The following seems to work but would need more tes…
-
The released stage 2 weight of resolution 256 seems to be incomplete, the error log is shown below.
`
File "/home/user/data/PT/PCDMs/stage2_batchtest_inpaint_model.py", line 126, in inference
…
-
# 🐛 Bug
I am currently experimenting with different scaled dot product attention implementations to evaluate training speed and GPU memory consumption.
I compared all methods running the followi…
-
**Describe the bug**
When using rotary positional embeddings and BF16, a type mismatch is produced:
```
Traceback (most recent call last):
File "/home/hatef.4/neox/gpt-neox/train.py", line 35, …
-
Modify the original attention
```
class Attention(nn.Module):
def __init__(self, args: ModelArgs):
super().__init__()
self.n_kv_heads = args.n_heads if args.n_kv_heads is None…
-
I am currently working with your TransNeXt model, specifically the tiny variant. When I use an image size that is not 224 (e.g., 138), I encounter an error in the attention_native.py file. The specifi…
-
## 🐛 Bug
### To Reproduce
Steps to reproduce the behavior (**always include the command you ran**):
1. Run cmd
fairseq-train $DATASET_DIR --arch $BART --restore-file $PRETRAINED_MODEL …
ghost updated
2 years ago
-
Hello,
I am not sure If I fully understand how to select the value of k for getting accurate results. Can you please guide me on which factors should be considered by deciding the k value.
Thank…
-
We need to have a way to deal with position errors in AARTFAAC when using pyse.py/TraP. There are two options for this:
- convert extremes of the positions to position error
- assume zero error to beg…
-
## TL; DR
- ViT feature representations are *less hierarchical*.
- Early tr blocks learn both local and global dependencies provided with large enough dataset.
- Skip connections play much more i…