菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-03
📄 Abstract - DALI: A Workload-Aware Offloading Framework for Efficient MoE Inference on Local PCs

Mixture of Experts (MoE) architectures significantly enhance the capacity of LLMs without proportional increases in computation, but at the cost of a vast parameter size. Offloading MoE expert parameters to host memory and leveraging both CPU and GPU computation has recently emerged as a promising direction to support such models on resourceconstrained local PC platforms. While promising, we notice that existing approaches mismatch the dynamic nature of expert workloads, which leads to three fundamental inefficiencies: (1) Static expert assignment causes severe CPUGPU load imbalance, underutilizing CPU and GPU resources; (2) Existing prefetching techniques fail to accurately predict high-workload experts, leading to costly inaccurate prefetches; (3) GPU cache policies neglect workload dynamics, resulting in poor hit rates and limited effectiveness. To address these challenges, we propose DALI, a workloaDAware offLoadIng framework for efficient MoE inference on local PCs. To fully utilize hardware resources, DALI first dynamically assigns experts to CPU or GPU by modeling assignment as a 0-1 integer optimization problem and solving it efficiently using a Greedy Assignment strategy at runtime. To improve prefetching accuracy, we develop a Residual-Based Prefetching method leveraging inter-layer residual information to accurately predict high-workload experts. Additionally, we introduce a Workload-Aware Cache Replacement policy that exploits temporal correlation in expert activations to improve GPU cache efficiency. By evaluating across various MoE models and settings, DALI achieves significant speedups in the both prefill and decoding phases over the state-of-the-art offloading frameworks.

顶级标签: llm systems model training
详细标签: mixture of experts inference optimization parameter offloading cpu-gpu co-execution workload-aware caching 或 搜索:

DALI:一种面向本地PC高效MoE推理的工作负载感知卸载框架 / DALI: A Workload-Aware Offloading Framework for Efficient MoE Inference on Local PCs


1️⃣ 一句话总结

本文提出了一种名为DALI的智能框架,它通过动态分配计算任务、精准预测和优化缓存策略,显著提升了在普通个人电脑上运行大型混合专家语言模型的效率和速度。

源自 arXiv: 2602.03495