LMD0311 / PointMamba

[NeurIPS 2024] PointMamba: A Simple State Space Model for Point Cloud Analysis
https://arxiv.org/abs/2402.10739
Apache License 2.0
361 stars 24 forks source link

Anybody observe slow training speed for mamba compared to transformer model? #14

Open vgthengane opened 6 months ago

d62lu commented 4 months ago

yes, I just made the comparison. The speed of the Mamba block is truly slower than Transformer, under the same input dimension

d62lu commented 4 months ago

for both model training and inference. I am not looking into the details of the Mamba block, maybe I missed something in the code....