naamiinepal / xrayto3D-benchmark

GNU General Public License v3.0
9 stars 3 forks source link

Explore: Efficient Transformers #20

Open msrepo opened 1 year ago

msrepo commented 1 year ago

Description

Currently, we are struggling to train transformer models for inputs larger than 128^3 volume. Within the volumetric grid representation of shape, the resolution achievable from this may not be enough. For example, for large bone shapes such as hip and ribs, we have had to use voxel resolution of ~2.5mm which is very coarse. We might be able to output a more high resolution image if the transformers were lighter in terms of memory usage.

Proposal

1) Use a 2D encoder and a 3D decoder. How do we execute this using transformer? A case in point how do we port the ideas from TransVert, 3DReconNet into pure transformers. How do we concatenate features from the two parallel x-ray branches and decode a 3D output? 2) Implement/port Conv+transformer These might be more efficient in terms of memory usage. Try: 1) implement/port various lightweight transformers and other interesting ideas