Published: September 17, 2024
Author(s)
Krishna Khadka (UTA), Sunny Shree (UTA), Yu Lei (UTA), Richard Kuhn (NIST), Raghu Kacker (NIST)
Conference
Name: International Workshop on Combinatorial Testing (IWCT) 2024
Dates: 05/27/2024 - 05/31/2024
Location: Toronto, Canada
Citation: 2024 IEEE International Conference on Software Testing, Verification and Validation Workshops (ICSTW), pp. 180-187
Machine Learning (ML) models rely on capturing important feature interactions to generate predictions. This study is focused on validating the hypothesis that model predictions often depend on interactions involving only a few features. This hypothesis is inspired by t-way combinatorial testing for software systems. In our study, we utilize the notion of Shapley Additive Explanations (SHAP) values to quantify each feature’s contribution to model prediction. We then use a greedy approach to identify a minimal subset of features (t) required to determine a model prediction. Our empirical evaluation is performed on three datasets: Adult Income, Mushroom, and Breast Cancer, and three classification models: Logistic Regression, XGBoost, and SVM. Through our experiments, we find that the majority of predictions are determined by interactions involving only a subset of features.
Machine Learning (ML) models rely on capturing important feature interactions to generate predictions. This study is focused on validating the hypothesis that model predictions often depend on interactions involving only a few features. This hypothesis is inspired by t-way combinatorial testing for...
See full abstract
Machine Learning (ML) models rely on capturing important feature interactions to generate predictions. This study is focused on validating the hypothesis that model predictions often depend on interactions involving only a few features. This hypothesis is inspired by t-way combinatorial testing for software systems. In our study, we utilize the notion of Shapley Additive Explanations (SHAP) values to quantify each feature’s contribution to model prediction. We then use a greedy approach to identify a minimal subset of features (t) required to determine a model prediction. Our empirical evaluation is performed on three datasets: Adult Income, Mushroom, and Breast Cancer, and three classification models: Logistic Regression, XGBoost, and SVM. Through our experiments, we find that the majority of predictions are determined by interactions involving only a subset of features.
Hide full abstract
Keywords
combinatorial testing; feature interaction; model prediction
Control Families
None selected