-
Thank you ematvey for this paper.
I wonder the uw and us are two vectors as global weights, or there are different uw(s) for each sentence, and different us(s) for each document?
From the code I…
-
### Description
With #1523, it's easy to train models similar to UGraphEmb. The UGraphEmb model structure is interesting, using attention and multiple layers.
1. Bai Y, Ding H, Qiao Y, et al. Un…
huonw updated
4 years ago
-
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the…
-
`sdpa_ex` implementation of `torch.nn.functional.scaled_dot_product_attention` returns all output tensor proxy in trace to be on `cuda` but at runtime some outputs are on `cpu`.
Repro
```python
i…
-
A matomo user wants to save graphs as SVG. However, the only option available is PNG.
![image](https://github.com/matomo-org/matomo/assets/132400623/b017dfd1-a4fe-4497-b402-b36ea89198dc)
![image](…
-
Hello Ma,
I am trying to run your code but it requires
- 'allx',
- 'ally',
- 'graph',
- "adjmat",
- "trainMask",
- "valMask",
- "testMask"
I checked the preprocessing files and …
-
Hi author,
It's a great work! I tried to reproduce your result on WLASL dataset. But I found that the trained result is bad (top1 per instance = 1.66%, top5 per instance= 1.77%) using your preproce…
pcc03 updated
3 months ago
-
I'm proposing to consider a new type of SHACL Shape (e.g. 'RepositoryShape') to enforce a vendor-neutral constaints on RDF graphs on a repository level (or triplestore-level in case if there is to 're…
-
When I run:
python llm_export.py --type Qwen-7B-Chat --path /mnt/LLM_Data/Qwen-7B-Chat --export_split --export_token --export_mnn --onnx_path /mnt/LLM_Data/Qwen-7B-Chat-onnx --mnn_path /mnt/LLM_Data/…
-
If pad tokens are used, and `model.eval(); model.train()` is called, Unsloth backward pass is undifferentiable, resulting in `nan`.
Reproduction script (expand):
```
import torch
from transf…
lapp0 updated
1 month ago