Closed dennisushi closed 8 months ago
My mistake is I am passing -l language_features
instead of -l language_features_dim3
, because those were not generated during autoencoder training. I didn't realize we have to use python test.py --dataset_path ../$DATASET_PATH --dataset_name $DATASET_NAME
to downsample the dimensions to 3.
language_feature = torch.zeros((self._xyz.shape[0], 3), device="cuda")
from hereWhy are language features set to 3 dimensions? In the L1 loss computation then we have (3,H,W) predicted features and (512,H,W) precomputed GT features. Is this normal / supposed to be a tunable parameter as in Table 7 of the paper?