-
### Model Name
MolE molecular embeddings
### Model Description
MolE is a foundation model for chemistry developed by Recursion. It combines geometric deep learning with transformers, to learn a mea…
-
Original Repository: https://github.com/ml-explore/mlx-examples/
Listing out examples from there which would be nice to have. We don't expect the models to work out the moment they are translated to …
-
1.(HMMR) Learning 3d human dynamics from video(2019)
temporal encoder: **1D temporal** convolutional layers, **precompute** the image features on each frame, get current and ±∆t frames prediction.
c…
-
# New Operator
Self Attention
### Describe the operator
Multi-headed attention is seeing prolific use in all transformers (mostly described in [pytorch](https://pytorch.org/docs/stable/generated/t…
dfiru updated
3 weeks ago
-
1.PARE: Part Attention Regressor for 3D Human Body Estimation(2021)
img-->volumetric features(before the global average pooling)-->part branch: estimates attention weights +feature branch: performs S…
-
I do not have a clear idea about how Hivemind should be integrated, yet. Let this issue exist to document a discussion around potential solutions.
Currently, each layer in a decoder is called an "e…
-
感谢作者在可解释性方面做出的优秀工作。我目前在用llama-3.1-8b做一些研究,在给transformer_lens中添加了`meta-llama/Llama-3.1-8B-Instruct`的支持代码后,运行`knowledge_eap.ipynb`,发现在第6个单元格计算`attribute(model, g, data, partial(logit_diff, loss=True, me…
-
### 🐛 Describe the bug
Was doing accuracy aware quantization for the model Marqo/marqo-fashionSigLIP . It was working for more than 2 hours+ and suddenly generated error. It was also not converging v…
-
pip install git+https://github.com/NVIDIA/TransformerEngine.git@stable
Thanks in advance.
---------------------------------------
python: 3.9
pytorch: 2.1.0
cuda: 12.1
gcc: 7.5.0
FAILED: CMakeFiles…
-
### System Info
```Shell
- `Accelerate` version: 0.34.2
- Platform: Linux-5.14.0-362.24.2.el9_3.x86_64-x86_64-Intel-R-_Xeon-R-_Silver_4216_CPU_@_2.10GHz-with-glibc2.37
- `accelerate` bash locati…