Closed boche closed 8 years ago
Hi, @boche Have you solved the problem?
When you split your data into 2 parts using two machines, I assume you would like to train them parallel. Then each machine should contain only one block, with same name block.0
.
The argument -num_blocks
need to be 1.
I'm not sure whether I talked is your setting.
Indeed, you are right, I noticed it after reading code.
Hi, so I can run lightlda in single machine, but met problem when I tried to run them in distributed mode (mpi).
Here is my steps:
and distribute them into two machines.
and eventually failed with information:
Any idea why it happens?