Abstract Interpretation-Based Feature Importance for Support Vector Machines

Abstract

We study how a symbolic representation for support vector machines (SVMs) specified by means of abstract interpretation can be exploited for: (1) enhancing the interpretability of SVMs through a novel feature importance measure, called abstract feature importance (AFI), that does not depend in any way on a given dataset or the accuracy of the SVM and is very fast to compute; and (2) certifying individual fairness of SVMs and producing concrete counterexamples when this verification fails. We implemented our methodology and we empirically showed its effectiveness on SVMs based on linear and nonlinear (polynomial and radial basis function) kernels. Our experimental results prove that, independently of the accuracy of the SVM, our AFI measure correlates much strongly with stability of the SVM to feature perturbations than major feature importance measures available in machine learning software such as permutation feature importance, therefore providing better insight into the trustworthiness of SVMs.

Publication
In Proc. 25th International Conference on Verification, Model Checking, and Abstract Interpretation (VMCAI 2024)
Awarded the Functional, Reusable, and Available Artifact Evaluation Badges
Acceptance: 40.5%
Date

Libra