-
Hi there, I've been following this work for a few months and found it's really an amazing idea to run LLMs over the Internet, while I'm also trying to improve Petals' performance on model inference in…
-
Click to expand!
### Issue Type
Bug
### Source
binary
### Tensorflow Version
tf 2.10
### Custom Code
Yes
### OS Platform and Distribution
Ubuntu 18.04
### Mobile de…
-
Thanks for the great work! Are there any tips for training with the improved contrastive divergence objective? I'm trying to build a multi-modal robotic manipulation model that takes in videos and tex…
-
Hello, thank you for your work.
I have interest in the AttentiveFP implementation from the paper "Pushing the Boundaries of Molecular Representation for Drug Discovery with the Graph Attention Mech…
-
### Feature request
The current implementation of the LLAMA model in the Hugging Face Transformers repository supports self-attention layers as per the standard design of transformer models. I prop…
-
Overall, Beyond Tech testnet demonstrates a promising foundation for EVM swaps, but there are several technical areas that i think might require attention to optimize performance and user experience.
…
-
### 🚀 The feature, motivation and pitch
1. NotImplementedError: Could not run 'aten::_to_copy' with arguments from the 'NestedTensorXPU' backend
cases:
test_transformers.py::TestTransformersXPU::te…
-
**Issue: Multi-Head Attention Producing Incorrect Vectors**
The multi-head attention mechanism in our transformer model appears to be producing incorrect vectors. Specifically, the attention matrix…
-
### System Info
```shell
Collecting environment information...
WARNING 11-10 14:19:08 _custom_ops.py:14] Failed to import from vllm._C with ImportError('/mnt/bbuf/vllm-backup/vllm/_C.abi3.so: undef…
-
Hello, I am working on a 2-dimensional UNet for sparse image denoising and I would like to integrate the attention mechanism (below) into the UNet network. But I did not succeed, if possible, I would …