-
With #168 this is surprisingly close, besides all the things that GeoData does that xarray can't do.
But clearly missing are:
- [ ] Dask-like processing for larger-than-memory files. This should a…
-
> One should never rely on the number of bytes actually allocated corresponding to the number requested.
The number of bytes allocated is guaranteed to be the same (or more? I guess it's rounded up…
-
### What is the issue?
**Description:**
I encountered an issue where the **LLaMA 3.2 Vision 11b** model loads entirely in CPU RAM, without utilizing the GPU memory as expected. The issue occurs on m…
-
Hello, @Snosixtyboo @ameuleman my device is 4090 24G.
First,when using the SIBR viewer to view my trained model (model size is 4G), I found that the gpu memory is about 22G, if this is the case, if…
-
**Severity**: Medium
**Vulnerability Details**:
Even after fixing the dynamic size allocation, there is a bug where retData is still pre-allocated to a fixed size (2 * 32 bytes). This allocation s…
-
Hi, this is great news after 14 years!
i've tried the executable on an Epyc 7282 server 2019 with 128 GB RAM and on djvu files larger than ~ 25MB it gives for example:
Processing page 81 of 81
…
-
Hello, Thanks for this library; I like your approach to the problem. But I was testing and it seems to be using more GPU memory than I expected?
If I simply change the `grid_num` in the `examples/t…
jtbr updated
5 months ago
-
Hello Team,
It seems the MacOS version took 1.4G memory whenever open it. I imagine what happened, even with no tree loaded it constantly consumes 1.4G memory, even larger than famous software such…
-
Hello, I'm using davis346 whose resolution is higher than davis240. Therefore, the computation of motion compensation is larger. I find that the lag time becomes longer and longer during use, and the …
-
## Introduction
I propose to update the NVS filesystem code to have the NVS_BLOCK_SIZE configurable in its Kconfig.
### Problem description
The NVS_BLOCK_SIZE is currently is set as a pr…