RADE: resource-efficient supervised anomaly detection using decision tree-based ensemble methods

Shay Vargaftik, Isaac Keslassy, Ariel Orda, Yaniv Ben-Itzhak

Research output: Contribution to journalArticlepeer-review

7 Scopus citations

Abstract

The capability to perform anomaly detection in a resource-constrained setting, such as an edge device or a loaded server, is of increasing need due to emerging on-premises computation constraints as well as security, privacy and profitability reasons. Yet, the increasing size of datasets often results in current anomaly detection methods being too resource consuming, and in particular decision-tree based ensemble classifiers. To address this need, we present RADE—a new resource-efficient anomaly detection framework that augments standard decision-tree based ensemble classifiers to perform well in a resource constrained setting. The key idea behind RADE is first to train a small model that is sufficient to correctly classify the majority of the queries. Then, using only subsets of the training data, train expert models for these fewer harder cases where the small model is at high risk of making a classification mistake. We implement RADE as a scikit-learn classifier. Our evaluation indicates that RADE offers competitive anomaly detection capabilities as compared to standard methods while significantly improving memory footprint by up to 12 × , training-time by up to 20 × , and classification time by up to 16 ×.

Original languageEnglish
Pages (from-to)2835-2866
Number of pages32
JournalMachine Learning
Volume110
Issue number10
DOIs
StatePublished - Oct 2021

Keywords

  • Anomaly detection
  • Decision-tree based ensemble methods
  • Fast machine learning
  • Resource efficient machine learning
  • Supervised learning

ASJC Scopus subject areas

  • Software
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'RADE: resource-efficient supervised anomaly detection using decision tree-based ensemble methods'. Together they form a unique fingerprint.

Cite this