Shap :: Anaconda.org
https://anaconda.org/conda-forge/shapDescription. SHAP (SHapley Additive exPlanations) is a unified approach to explain the output of any machine learning model. SHAP connects game theory with local explanations, uniting several previous methods and representing the only possible consistent and locally accurate additive feature attribution method based on expectations.
No module named 'shap.explainers' with 0.13 in LightGBM ...
https://github.com/slundberg/shap/issues/5409.04.2018 · egg / shap / __init__. py in < module > () 5 from iml. links import Link, IdentityLink, LogitLink 6 from iml. common import Instance, Model----> 7 from. explainers. kernel import KernelExplainer, kmeans 8 from. explainers. tree import TreeExplainer 9 from. plots import visualize, plot, summary_plot, joint_plot, interaction_plot, dependence_plot, force_plot …
shap · PyPI
https://pypi.org/project/shap20.10.2021 · SHAP (SHapley Additive exPlanations) is a unified approach to explain the output of any machine learning model. SHAP connects game theory with local explanations, uniting several previous methods and representing the only possible consistent and locally accurate additive feature attribution method based on expectations.