Closed uriyapes closed 1 year ago
Hi @uriyapes , thanks for your interest in our work and good questions!
What I wish to know is what is the dimension of each hierarchical query q^{hie}_{ij}?
The dimension of each query is set to 256.
Is this done as described in the Deformable Attention paper?
Yes. We have not changed the intrinsic mechanism of Deformable Attention.
I got two questions:
“In each decoder layer, we adopt MHSA to make hierarchical queries exchange information with each other (both inter-instance and intra-instance)” So, inter-instance and intra-instance interaction in one MHSA layer? or two layers?
The initial quires are learnable params, right?
So, inter-instance and intra-instance interaction in one MHSA layer? or two layers?
We perform the interaction in one MHSA layer.
The initial quires are learnable params, right?
Yes, the initial queries are learnable embeddings.
And we release an initial version of MapTR, you can refer to the code for more details. I'm closing this issue, but let us know if you have further questions.
Hi, I've just read the paper and the idea and results are impressive. I know you haven't release the source code yet but since I want to present this paper in my workplace I would like to better understand the architecture of your network. I have two questions:
Many thanks in advance.