Open windar427 opened 5 years ago
Hi @windar427
thanks,
iloc
I find the other error in
if len(np.unique(y)) > 100 or len(np.unique(y)) > 0.1 * y.shape[0]:
print("regression")
print("meta features cannot be extracted as the target is not categorical")
# if classification
else:
# print("classification")
metafeatures_clf = {}
# compute clustering performance metafeatures
metafeatures_clf['silhouette'], metafeatures_clf['calinski_harabaz'], metafeatures_clf[
'davies_bouldin'] = compute_clustering_metafeatures(X)
# compute landmarking metafeatures
metafeatures_clf['naive_bayes'], metafeatures_clf['naive_bayes_time'] = pipeline(X, y, GaussianNB())
metafeatures_clf['linear_discriminant_analysis'], metafeatures_clf[
'linear_discriminant_analysis_time'] = pipeline(X, y,
LinearDiscriminantAnalysis(solver='lsqr', shrinkage='auto'))
metafeatures_clf['one_nearest_neighbor'], metafeatures_clf['one_nearest_neighbor_time'] = pipeline(X, y,
KNeighborsClassifier(
n_neighbors=1))
metafeatures_clf['decision_node'], metafeatures_clf['decision_node_time'] = pipeline(X, y,
DecisionTreeClassifier(
criterion='entropy',
splitter='best',
max_depth=1,
random_state=0))
metafeatures_clf['random_node'], metafeatures_clf['random_node_time'] = pipeline(X, y, DecisionTreeClassifier(
criterion='entropy', splitter='random', max_depth=1, random_state=0))
metafeatures = list(metafeatures_clf.values())
return metafeatures
if the regression task, metafeatures
referenced None
Yes, that should be the case. datacleanbot
can only be used to deal with supervised classification tasks for now. Classifications labels are required to compute metafeatures
so, that should be ignored or metafeatures = None
so, that should be ignored or
metafeatures = None
Ah yes. Thanks! I will update that.
in abda no module call abda.bin.spstd_model_ha1.py
Only a small part of the Bayesian model is used in datacleanbot
in
1.X shoud be X=Xy.iloc[:, :-1] and y =Xy.iloc[:, -1] 2.when the data was not clean, should we change the run sequence, put
show_important_features
to the end