Open FaridehJalali85 opened 3 years ago
MIToolbox operates on discrete inputs so you will need to discretise them before using it, otherwise it will apply a standard discretisation which probably doesn't do what you want. In the past we've used 10 bins of equal width in the range (min, max) and that has tended to work reasonably well.
I believe scikit-learn has a continuous/discrete mutual information calculation, or there are packages like ITE (https://bitbucket.org/szzoli/ite-in-python/src/master/) which provide many different estimators for the mutual information.
Thanks Adam
Can you please elaborate more about 10 bins of equal width in the range (min, max) for the outcome? Is there any other approach that we can map the continuous output to the classification task? Thanks Fari
There are many different binning algorithms. Equal frequency binning (where the bin widths are set to ensure each bin has the same number of elements in it) interacts oddly with information theoretic feature selection, as it makes each feature maximum entropy. We used equal width binning in our papers on feature selection and it worked well. You can also set the bins based on mean & std dev if you think the variable is approximately gaussian distributed, or use some meaningful bins if you have domain knowledge about the feature values.
Thanks Adam,
Can we have multi-label class(outcome) with your developed information theoretic feature selection toolbox or should it be only binary? Thanks
Multi-class is fine.
Hi Adam,
I have 500 examples, continues outcome ( above and below 0) and 33000 features. I would like to use feast,... for finding the most informative features. As far as I know, I have to discretise the outcome. Can you please advise how I should that? Also do you aware of any information theoretic packages that can deal with the continues outcome?
Please see some of numbers in the list for the outcome.
0.55228076 -0.3197724 0.58774863 -0.5174945 -0.173138 1.27375815 0.78408593 0.74372201 -1.3886196 0.43545769 -0.0689654 0.18626918 1.8202002 0.10355088 -0.0560193 0.29268956 -0.4401979 1.61399178 -0.704535 0.00430693 1.41157343 -0.0646488 -1.4474665 0.4307273 0.44970806 -0.3885697 1.10758465 -0.1339137 0.12522725 -0.3379575 -1.099666 0.31524279 -0.7666343 0.22144974 -0.8017797 -1.1909404 -0.7958541 -1.5830851 1.03128861 -1.0312886 -1.736507 0.77242535 1.06106625 0.95375794 -0.8626751 -0.4025419 -0.1078819 0.71006597 0.5422753 -2.2562499 -1.3238074 -1.919012 1.24519952 -0.4930275 0.47848871 1.21762188 -0.6239719 1.42333462 -0.0086139 0.89435119 -1.698852 1.77679865 -0.038772 -0.2569066 0.28371043 -0.360848 -0.2658198 0.8440787 -0.9812682 0.67719571 0.31977237 -2.0077154 0.48332361 0.70453501 0.52242483 -0.9537579 -0.8197281 -1.2835121 -0.5076713 0.35625516 0.03015302 0.05601926 -0.1994326 1.13176806 -2.3148972 1.3886196 0.5724601 -1.8925387 -1.1822326 -0.2926896 0.37467286 -0.6610331 -0.2971879 -1.3555018 -1.5536207 0.73805488 -0.3746729 -0.6031758 -0.6935372 1.43529606 1.48532767 0.629213 0.19943255 -0.7211937 -1.068659 -0.4165933 0.01292112 1.26412388 -0.9335899 0.22586574 -2.2044686 -0.6450414 -0.5724601 1.00954199 0.05170609 -1.0840322 -0.5027781 -0.7782424 -0.2258657 -0.5623418 -0.1644006 -0.7156187 -0.8879448 -0.2792295 -0.0776024 -0.4544784 -0.8137162 0.93358993 0.73241136 1.89253869 -1.366375 -1.472472 0.15131773 -0.6135407 -0.1165502 0.80177968 0.92695117 1.71737513 0.77824241 -1.4853277 -0.2347111 -0.5522808 0.85644338 0.09922176 -0.5573042 -1.1566322 0.34252111 0.37005675 1.51180219 -1.5118022 0.02153627 0.37929696 1.28351212 -0.0948945 -0.7551286 1.59834762 0.14696264 -0.4784887 1.56818275 0.60317579 0.82576977 1.05353422 0.33795753 0.83184174 -2.702943 -0.3107197 0.72679107 1.40000431 -0.379297 -0.2702843 -0.4307273 1.18223265 0.3516698 1.13997784 1.73650701 0.60835006 1.47247197 0.03446219 1.55362072 -0.5224248 1.06865897 2.46384107 0.38856968 -0.333401 0.18188866 2.31489724 -0.3654485 -0.5422753 1.36637495 -0.112215 -0.3152428 0.42600652 -0.1426103 0.94027009 -0.7899563 -0.5472712 0.39787592 1.25460554 -1.3135329 0.42129522 0.12088759 0.86267505 0.99530565 -1.4233346 0.97432155 0.19065327 0.26581978 -0.6664013 -2.158 0.9007945 0.0776024 -0.3016924 -1.9763958 0.27475411 1.08403222 0.81972809 0.23028616 2.00771543 1.33422361 0.55730423 -0.9402701 0.76086887 -0.6083501 0.14261033 -0.4354577 1.6634623 -1.009542 -0.473665 -0.6880696 0.34709182 -1.6634623 0.65035348 -0.3562552 0.92035301 0.85024479 -0.6717887 1.04606148 -0.4978969 0.4544784 -1.680894 -2.1157713 -0.1469626 1.94690278 -2.041138 0.49789692 -1.1650809 0.11655021 -1.1317681 1.03864671 -1.6300412 2.15799996 1.16508088 0.30169235 0.23471108 -2.5652793 0.01722854 -2.382787 0.46885224 -0.8318417 0.038772 2.92573583 0.41659325 0.41190048 0.58263735 -0.2747541 -1.6139918 -0.1687677 -0.0430825 0.78995634 0.40721675 -1.0918158 -1.1236339 -1.2176219 0.25690662 -1.5393809 -0.030153 -1.7562965 1.52544651 -0.9882626 -1.3774121 0.83794453 -0.6771957 0.24801382 -1.4115734 0.27028425 -1.7767986 0.56739369 -2.0770094 -1.2546055 -1.7173751 -0.9007945 -0.2126306 -0.1818887 -1.0763138 -0.3470918 -0.2038279 -1.2086323 0.07328321 0.20822723 -0.8564434 1.09966602 0.96742157 1.97639581 -0.2837104 -0.920353 0.71561873 1.91901198
Thanks Fari