Closed semi-supervised-paper closed 2 years ago
Hi - thanks for your interest. This is a great question. I'll split my thoughts into 2 parts.
First for LDS, since the underlying idea is to smooth over the label space, it can be straightforwardly extend to multi-dim targets using multivariate KDE (see an implementation using Gaussian kernel in Python here). Otherwise, the principle and implementation should be the same.
For FDS, it might be a bit more complex. Since it operates over the feature space, it depends on the size of your intermediate features. If a feature vector is available for each of the target dimension (i.e., [bsz, feat_dim, out_dim]
), FDS is directly applicable by smoothing features over each dimension. Otherwise, if only 1 vector available for one sample (i.e., [bsz, feat_dim]
), one straightforward way is to first determine the high-dimensional distance between samples (using multivariate KDE or other methods), then smooth the features using FDS with respect to such high-dim distance.
To sum up, for either technqiue, the idea is the same when extending to multi-dim outputs, and the extension should be straightforward. The key is still to estimate the "effective" label density distribution for multi-dimensional continuous labels.
Thanks for your suggestion ! I'll close the issue.
Hi @semi-supervised-paper, Have you been able to adapt LDS and/or FDS to multi-dimensional targets using multivariate KDE? My aim is to train a convolutional neural network to train on three continuous targets. I would be grateful if you could share the solution to your problem.
Hi, Amazing for your great job in the imblanced regression problem. But I notice that this work discusses more on the 1D regression problems. What if the output is more than 1D (like Batch_Size x 10) ? Any suggestion will be helpful. Thanks.