Xtra-Computing / thundergbm

ThunderGBM: Fast GBDTs and Random Forests on GPUs
Apache License 2.0
692 stars 87 forks source link

There was an error using thundergbm-predict file #14

Closed lvpinrui closed 5 years ago

lvpinrui commented 5 years ago

My environment is: Windows 10 Visual Studio 2017 Community CMake3.14.0-rc3 CUDA10.1.105 Using test_dataset.txt in the library is successful,but Using my own .txt is failed. my thundergbm-predict: image

When building thundergbm-predict : image

jiahuanluo commented 5 years ago

Dear, I have run several tests on similar dataset usps, and no error occurred. Could you please provide more details or zip your dataset and put it here if available? Make sure that your dataset is in a LibSVM format. Thanks.

lvpinrui commented 5 years ago

亲爱的, 我在类似的数据集usps上运行了几次测试,没有发生错误。 您能否提供更多详细信息或压缩您的数据集并将其放在此处(如果有)? 确保您的数据集采用LibSVM格式。 谢谢。 hello,this is my dataset: 1.txt

waiting for your advice,thanks!

jiahuanluo commented 5 years ago

Hi there, We found that your dataset is zero_based (the first index of feature is 0). The newest version is able to handle zero_based dataset. You can use the newest code or turn your dataset into one-based. Thanks.

lvpinrui commented 5 years ago

hello,Running the original code after changing my dataset succeeded.but i use the newest code,This has a new problem. my thundergbm-train: image my dataset is always 1.txt, When building thundergbm-train : image

waiting for your advice,thank you very much.

jiahuanluo commented 5 years ago

Hi lvpinrui,

Thanks for imforming. You can delete the code in this line and rebuild the project. There is a bug here, we will fix it in no time.

Regards

lvpinrui commented 5 years ago

i use the newest code,When my dataset is 1.txt running thundergbm-predict file succeeds, but when the dataset is 2.txt thundergbm-predict file fails ,this is a new problem image my dataset is 2.txt,2.txt file size is 1.43GB,I don't know if it is because the file is too big.

jiahuanluo commented 5 years ago

It should work. You should check your dataset, or share it through the cloud (Baidu Yun?)

lvpinrui commented 5 years ago

this my dataset(Baidu Yun): 链接:https://pan.baidu.com/s/15H6YW_mqZP6K4hrOVl9eOA 提取码:s9y7

thank you very much!

jiahuanluo commented 5 years ago

I have tested your dataset 2.txt on the newest code. No error has occurred. @lvpinrui

lvpinrui commented 5 years ago

我已在最新代码上测试了您的数据集2.txt。没有发生错误。 @lvpinrui

hello ,thank you very much for your help. but my Configuration is: max_depth=6 num_round=40 n_gpus=1 verbosity=0 profiling=0 data=C:\Users\Administrator\Desktop\2\2.txt max_bin=255 colsample=1 bagging=1 num_parallel_tree=1 eta=1.0 objective=multi:softmax num_class=8 min_child_weight=1 lambda=1 gamma=1 tree_method=auto
when building thundergbm-train is ok, when building thundergbm-predict is failed.

jiahuanluo commented 5 years ago

Yes, the same configuration as yours. I ran the test via command line, editing the machine.conf. It works fine for me. @lvpinrui

lvpinrui commented 5 years ago

Yes, the same configuration as yours. I ran the test via command line, editing the machine.conf. It works fine for me. @lvpinrui

thank you very much for your help these days. I tried it a few times or not, I succeeded after changing my dataset.

jiahuanluo commented 5 years ago

Since the issue is solved, I will close it.

Cheers, Locke