ArXiv TLDR

Quantifying Sensitivity for Tree Ensembles: A symbolic and compositional approach

🐦 Tweet
2605.13830

S. Akshay, Chaitanya Garg, Ashutosh Gupta, Kuldeep S. Meel, Ajinkya Naik

cs.AIcs.LG

TLDR

This paper introduces a novel symbolic and compositional method to quantify sensitivity in decision tree ensembles, efficiently identifying misclassification risks.

Key contributions

  • Proposes a quantitative sensitivity notion for DTEs by discretizing input space and enumerating susceptible regions.
  • Introduces a novel, efficient algorithmic technique with certified error and confidence bounds.
  • Encodes the problem using Algebraic Decision Diagrams (ADDs) for compositional and scalable computation.
  • Demonstrates significant speedup and scalability with tool XCount compared to model counters.

Why it matters

Decision tree ensembles are crucial in safety-critical AI, making sensitivity verification vital for reliability. This paper offers an efficient, scalable method to quantify sensitivity, addressing a key challenge in model verification.

Original Abstract

Decision tree ensembles (DTE) are a popular model for a wide range of AI classification tasks, used in multiple safety critical domains, and hence verifying properties on these models has been an active topic of study over the last decade. One such verification question is the problem of sensitivity, which asks, given a DTE, whether a small change in subset of features can lead to misclassification of the input. In this work, our focus is to build a quantitative notion of sensitivity, tailored to DTEs, by discretizing the input space of the model and enumerating the regions which are susceptible to sensitivity. We propose a novel algorithmic technique that can perform this computation efficiently, within a certified error and confidence bound. Our approach is based on encoding the problem as an algebraic decision diagram (ADD), and further splitting it into subproblems that can be solved efficiently and make the computation compositional and scalable. We evaluate the performance of our technique over benchmarks of varying size in terms of number of trees and depth, comparing it against the performance of model counters over the same problem encoding. Experimental results show that our tool XCount achieves significant speedup over other approaches and can scale well with the increasing sizes of the ensembles.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.