Closed Gabomfim closed 2 years ago
could you share your data set?
Sorry for keeping you waiting.
I'm sharing with you my notebook with all the files, including the databases used (in the data
file).
I managed to fix the problem by importing the database as txt instead of csv.
allstroke.txt
is the txt version of the healthcare-dataset-stroke-data.csv
database. That did the fix.
We now import the database in this way:
df = pd.read_csv("./data/allStroke.txt", index_col=0)
I don't have the old code with me now, but I can send it to you the next week if needed.
When I run this in my environment, it works well. I have Python 3.8.12, pandas==1.3.5. I recommend you to upgrade or downgrade to my environment level.
from chefboost import Chefboost as chef
import pandas as pd
df = pd.read_csv("healthcare-dataset-stroke-data.csv", index_col=0)
print(df.head())
configGBM = {'algorithm': 'C4.5', 'enableGBM': True, 'epochs': 7, 'learning_rate': 1, 'max_depth': 5, 'enableParallelism': False}
modelGBM = chef.fit(df = df, config = configGBM)
Accuracy: 82.00389105058366 % on 1028 instances Labels: ['Yes' 'No'] Confusion matrix: [[99, 35], [150, 744]] Precision: 73.8806 %, Recall: 39.759 %, F1: 51.6971 %
Tried to do the following on a dataset with float samples. (Running on Python 3.7)
Error Log: