Closed Martingales closed 4 years ago
The feature you want is not implemented. However, an alternative here may be ok for the similar purpose.
When I understand the code correctly, you re-assign half the samples in each class to be 0 (aka no class). And then init.EM
is able to deal with K larger than the current number of clusters?
How are the cluster assignments dealt with internally? Are they a must have assignment kept until the end, are they defining a start clustering or are they acting as a prior all the time?
What you are trying to do is not allowed under the model that mixture models based clustering is built on. In other words, our model assumptions are not general enough. However, the problem is interesting though not completely clear and we will have to think about it. Are you interested in communicating outside git so that we may have a conversation. However, I am not sure about what your suggestions will entail so will also be cautious about how much effort will be required.
That sounds great! As I can't see your email, can you contact me via my public github profile?
I know that implementing generalised models isn't easy. At this stage it was more important for me to know whether it was a bug or a genuine limitation.
Thank you for the clarification!
Hi everyone,
I use
init.EM
for semi-supervised clustering and I found that runninginit.EM
breaks R under the following circumstances:Background The reason I want to use
init.EM
is that I suspect I have subclusters in each of my existing clusters. But I also want to allow samples from other clusters to join a subcluster. In essence, I want to use my existing labels as prior information only but not determine the final clustering. That's why I would like to have K larger than the number of existing clusters.Error should be reproducible with the dev's own example:
sessionInfo()