-
with Llama-3.1-7B
load the model:
```txt
==((====))== Unsloth 2024.10.0: Fast Llama patching. Transformers = 4.44.2.
\\ /| GPU: NVIDIA GeForce RTX 3090. Max memory: 23.7 GB. Platform …
-
### Description
I am trying to pass `PRNGKey`s to a function, which is integrated by `odeint`.
Here is a simplified example reproducing the problem:
```python
from functools import partial
im…
-
Bumps [transformers](https://github.com/huggingface/transformers) from 4.36.1 to 4.38.2.
Release notes
Sourced from transformers's releases.
v4.38.2
Fix backward compatibility issues with Llama and …
-
I'm encountering a KeyError when trying to train Phi-3 using the unsloth library. The error occurs during the generation step with model.generate. Below are the details of the code and the error trace…
-
Hello :)
We used the default Unsloth Colab Pipeline to ft a LLAMA3.1 8B and replicated this as a notebook on an azure environment.
https://colab.research.google.com/drive/1Ys44kVvmeZtnICzWz0xgpRn…
-
### 🐛 Describe the bug
when I set the `dropout_p=0.0`, the result is different. But `dropout_p=-1`, the result is same. Maybe the op scaled_dot_product_attention has some bug. Please fix it, thank…
-
Tisane currently provides two types of conceptual relationships: `causes` and `associates_with`. This doc covers when and how to use these verbs.
If a user provides associates_with, we walk them t…
emjun updated
2 years ago
-
Hi,
I noticed that for certain sequence lengths, given the same inputs, FlexAttention's output differs from FlashAttention's output.
For example, with an input sequence of length 137, FlexAttent…
-
Hi, I have a question regarding the image generation process, specifically the `generate_image` function at https://github.com/baaivision/Emu/blob/main/models/modeling_emu.py#L185
According to this…
-
Encoded latents shape: torch.Size([1, 4, 16, 60, 90])
Requested to load SD3ClipModel_
Loading 1 new model
loaded completely 0.0 4541.693359375 True
Temporal tiling and context schedule disabled
…