菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-13
📄 Abstract - Active Imitation Learning for Thermal- and Kernel-Aware LFM Inference on 3D S-NUCA Many-Cores

Large Foundation Model (LFM) inference is both memory- and compute-intensive, traditionally relying on GPUs. However, the limited availability and high cost have motivated the adoption of high-performance general-purpose CPUs, especially emerging 3D-stacked Static Non-Uniform Cache Architecture (3D S-NUCA) systems. These architectures offer enhanced bandwidth and locality but suffer from severe thermal challenges and uneven cache latencies due to 3D Networks-on-Chip (NoC). Optimal management of thread migration and V/f scaling is non-trivial due to LFM kernel diversity and system heterogeneity. Existing thermal management approaches often rely on oversimplified analytical models and lack adaptability. We propose AILFM, an Active Imitation Learning (AIL)-based scheduling framework that learns near-optimal thermal-aware scheduling policies from Oracle demonstrations with minimal run-time overhead. AILFM accounts for both core-level performance heterogeneity and kernel-specific behavior in LFMs to maintain thermal safety while maximizing performance. Extensive experiments show that AILFM outperforms state-of-the-art baselines and generalizes well across diverse LFM workloads.

顶级标签: systems model training machine learning
详细标签: imitation learning thermal management cache architecture scheduling foundation model inference 或 搜索:

面向3D S-NUCA众核系统热感知与内核感知大模型推理的主动模仿学习 / Active Imitation Learning for Thermal- and Kernel-Aware LFM Inference on 3D S-NUCA Many-Cores


1️⃣ 一句话总结

本文提出了一种名为AILFM的智能调度框架,它利用主动模仿学习技术,在新型3D堆叠CPU上自动学习如何高效调度大模型的计算任务,既能有效控制芯片发热、防止过热,又能充分利用硬件性能,从而替代昂贵的GPU进行大模型推理。

源自 arXiv: 2604.11948