菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-17
📄 Abstract - Benchmarking IoT Time-Series AD with Event-Level Augmentations

Anomaly detection (AD) for safety-critical IoT time series should be judged at the event level: reliability and earliness under realistic perturbations. Yet many studies still emphasize point-level results on curated base datasets, limiting value for model selection in practice. We introduce an evaluation protocol with unified event-level augmentations that simulate real-world issues: calibrated sensor dropout, linear and log drift, additive noise, and window shifts. We also perform sensor-level probing via mask-as-missing zeroing with per-channel influence estimation to support root-cause analysis. We evaluate 14 representative models on five public anomaly datasets (SWaT, WADI, SMD, SKAB, TEP) and two industrial datasets (steam turbine, nuclear turbogenerator) using unified splits and event aggregation. There is no universal winner: graph-structured models transfer best under dropout and long events (e.g., on SWaT under additive noise F1 drops 0.804->0.677 for a graph autoencoder, 0.759->0.680 for a graph-attention variant, and 0.762->0.756 for a hybrid graph attention model); density/flow models work well on clean stationary plants but can be fragile to monotone drift; spectral CNNs lead when periodicity is strong; reconstruction autoencoders become competitive after basic sensor vetting; predictive/hybrid dynamics help when faults break temporal dependencies but remain window-sensitive. The protocol also informs design choices: on SWaT under log drift, replacing normalizing flows with Gaussian density reduces high-stress F1 from ~0.75 to ~0.57, and fixing a learned DAG gives a small clean-set gain (~0.5-1.0 points) but increases drift sensitivity by ~8x.

顶级标签: systems model evaluation benchmark
详细标签: anomaly detection time series iot evaluation protocol event-level augmentation 或 搜索:

基于事件级数据增强的物联网时序异常检测基准测试 / Benchmarking IoT Time-Series AD with Event-Level Augmentations


1️⃣ 一句话总结

这篇论文提出了一套新的评估方法,通过模拟传感器故障、数据漂移等真实扰动来测试物联网时序异常检测模型在事件层面的表现,发现没有一种模型能适应所有情况,并据此为不同场景下的模型选择提供了具体指导。

源自 arXiv: 2602.15457