-
### Description
The goal of this roadmap item is to improve the experience for customers who pay with STORJ tokens rather than a credit card. We aim to automate the process of upgrading to a "pro acc…
-
Dear sir,
I deployed a bot-executr contract on Arbitrum using your code. Currently, I want to extract the balance of the weth, however, I encountered this error.
ValueError: Execution revert…
-
### Description
#2247 better isolates the `AttributeRemoverTests` to just `AttributeRemover`. Along the way, a new assert helper was created that, unlike `assertMacroExpansion`, doesn't trim newlines…
-
## How to reproduce
Using a p4d.24xlarge:
```python
from parallelformers import parallelize
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "facebook…
-
Hi,
Firstly, thanks for the awesome work!
I want to use KTO with a quantized Mistral model but am getting pickle errors from the multiprocessing thread, probably since that changes the Embeddin…
-
Hello,
I'm trying to visualize the attention weights for Llama 3 when it generates text, but I am facing some complications. I slightly modified the Attention class to output the `scores` variable …
-
The lottery logic currently relies on a ERC721 token to handle NFT tickets.
https://bscscan.com/token/0x5e74094cd416f55179dbd0e45b1a8ed030e396a1
Owing to technical limitations (no batch minting…
-
Hi,
I tried to upgrade the python package and run v0.5, but again it's showing the following error:
TypeError Traceback (most recent call last)
in
10 # S…
-
Hello
in this part of your code in bertcrf class (forward fn), you write it is for pass the first token but i don't understand how this hanpped (and the len of seq_logist and seq_lables doed not chan…
-
Hi,
I noticed that for certain sequence lengths, given the same inputs, FlexAttention's output differs from FlashAttention's output.
For example, with an input sequence of length 137, FlexAttent…