I am having an issue where TabDDPM does really well when the total number of variables are < 10-15, but for modest to high numbers of binary categorical variables (even over 100, although my goal is 1000s), the loss (both mloss and gloss) quickly go to nan. This doesn't seem to be an issue with my dataset - I get the same behavior when I make a synthetic dataset of gaussian continuous variables and Bernoulli(p=.5) binary variables. Any help would be appreciated.
I am having an issue where TabDDPM does really well when the total number of variables are < 10-15, but for modest to high numbers of binary categorical variables (even over 100, although my goal is 1000s), the loss (both mloss and gloss) quickly go to nan. This doesn't seem to be an issue with my dataset - I get the same behavior when I make a synthetic dataset of gaussian continuous variables and Bernoulli(p=.5) binary variables. Any help would be appreciated.