Termination is based on no. of iterations, they can be increased to get better results for less no of terms overlap or they can be reduced to make it faster.
On HPC, it took like 4 hours to generate each graph.
1% of full 2gb wikipedia (preprocessed and top 5% tfidf)
Different lines are variations with F, CR and pop size.
Points to ponder
Untuned results may not be accurate, as we did 1 run to find the scores. The shuffled row order input may produce better scores and % improvement can be less. This problem is with all the previous results and all the datasets. This can be solved by selecting the mean score / the maximum score after doing multiple runs.
How to read graphs?
Results
F = 0.3, CR = 0.7, Pop = 10
F = 0.7, CR = 0.3, Pop = 10
F = 0.3, CR = 0.7, Pop = 30
F = 0.7, CR = 0.3, Pop = 30
Conclusion