-
### 🐛 Describe the bug
here a 'simple' way to reproduce the issue
```
import torch
@torch.jit.script
def snake(x, alpha):
x = x + alpha + 1e-9
return x
class Snake1d(torch.nn.Modul…
-
Hi, I have some exotic, but really simple linear algebra package that performs the so-called Face splitting product of two matrices and a vector.
Tha matrices A and B have the same number of rows, b…
-
**I'm sorry for bothering you guys, but I'm have hard time trying to create a global variable on go-lua of type table.**
I'd not any issue to return a table from a function call (_GO Lang internal …
-
# Số ca mắc Covid-19 - Luyện Code Online
[https://luyencode.net/problem/CORONA](https://luyencode.net/problem/CORONA)
-
### 🐛 Describe the bug
If I fuse the backward with compiled forward+loss, there's a higher peak memory than if I separate the backward from compiled forward+loss. It looks like the logits aren't bein…
ad8e updated
5 months ago
-
### 🐛 Describe the bug
```
import torch
ln = torch.nn.LayerNorm( 16)
t = torch.nested.nested_tensor( [torch.rand( (2, 8, 16)) for _ in range(4) ], layout=torch.jagged)
print( t.shape) # torch.S…
-
**[T](#title)itle**: refactor(core): replace `wrapped_type_t` with `std::indirectly_readable_traits`.
**[D](#desc)escription**:
See and
https://github.com/mpusz/mp-units/blob/89bafed96…
-
**Issue by [gbiele](https://github.com/gbiele)**
_Wednesday Oct 14, 2015 at 09:50 GMT_
_Originally opened as https://github.com/stan-dev/stan/issues/1641_
----
## Feature Request
**Category**
New …
-
I have a test harness which uses `rwkv.cpp` or `llama.cpp`. Either way, when it loads, I get this error:
```
$ result/bin/agent.py rwkv models/rwkv-pile-14b.bin
~ Initializing adapter: rwkv
Ille…
-
## Issue description
When I am running distributed and I simply set `CUDA_VISIBLE_DEVICES` in each rank:
- Running `torch.distributed.barrier()` makes rank 1 occupy GPU memory on the GPU of rank 0…