Closed lucasjinreal closed 2 years ago
Hi, the model checkpoint contains three subnetworks for body, hand, and face reconstruction. Moreover, each subnetwork also contains a self-attention module (which includes lots of parameters).
@HongwenZhang does the attention module using bert as encoder?
yes
I look inside the code, seems didn't use transformer, why the ckpt so big?