8000 GitHub - jimmafeni/EFI-Toolbox: A toolbox to produce more robust and accurate interpretations of feature importance in Python
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

jimmafeni/EFI-Toolbox

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EFI Toolkit

With the widespread use of machine learning to support decision-making, it is increasingly important to verify and understand the reasons why a particular output is produced. Although post-training feature importance approaches assist this interpretation, there is an overall lack of consensus regarding how feature importance should be quantified, making explanations of model predictions unreliable. In addition, many of these explanations depend on the specific machine learning approach employed and on the subset of data used when calculating feature importance.

A possible solution to improve the reliability and understandability of explanations is to combine results from multiple feature importance quantifiers from different machine learning approaches coupled with data re-sampling.

EFI toolkit implements this solution using:

  • State-of-the-art information fusion techniques
  • Fuzzy sets

The toolkit provides complete automation of the entire feature importance computation cycle.

INPUT = Structured data

The main attributes of the toolbox are:

  • automatic training and optimisation of machine learning algorithms.
  • automatic computation of a set of feature importance coefficients from ensemble of optimised machine learning algorithms and feature importance calculation techniques.
  • automatic aggregation of importance coefficients using multiple decision fusion techniques.
  • automatic generation of fuzzy membership functions that show the importance of each feature to the prediction task in terms of low', moderate' and `high' importance as well as their levels of uncertainty.

alt text

alt text

Dependencies

  • pip install dataframe-image
  • pip install scikit-learn
  • pip install tensorflow
  • pip install silence-tensorflow

Relevant publications

If you use EFI in scientific publications, we would appreciate citations.

EFI: A Toolbox for Feature Importance Fusion and Interpretation in Python Aayush Kumar, Jimiama Mafeni Mase, Divish Rengasamy, Benjamin Rothwell, Mercedes Torres Torres, David A. Winkler, Grazziela P. Figueredo Conference on Machine Learning, Optimization, and Data Science (2022)

Link to publication.

@article{kumar2022efi,
  title={EFI: A Toolbox for Feature Importance Fusion and Interpretation in Python},
  author={Kumar, Aayush and Mase, Jimiama Mafeni and Rengasamy, Divish and Rothwell, Benjamin and Torres, Mercedes Torres and Winkler, David A and Figueredo, Grazziela P},
  journal={arXiv preprint arXiv:2208.04343},
  year={2022}
}

Feature importance in machine learning models: A fuzzy information fusion approach Divish Rengasamy,Jimiama M.Mase, Aayush Kumar, Benjamin Rothwell, Mercedes Torres Torres, Morgan R.Alexander, David A.Winkler, Grazziela P.Figueredo* Neurocomputing Journal, 2022

Link to publication.

@article{rengasamy2022feature,
  title={Feature importance in machine learning models: A fuzzy information fusion approach},
  author={Rengasamy, Divish and Mase, Jimiama M and Kumar, Aayush and Rothwell, Benjamin and Torres, Mercedes Torres and Alexander, Morgan R and Winkler, David A and Figueredo, Grazziela P},
  journal={Neurocomputing},
  year={2022},
  publisher={Elsevier}
}


**The toolkit and its description will be updated as new explainability techniques and machine learning models are implemented.**

About

A toolbox to produce more robust and accurate interpretations of feature importance in Python

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published
0