这太耗时了——探究将时间作为大语言模型能耗的代理指标 / This Is Taking Too Long -- Investigating Time as a Proxy for Energy Consumption of LLMs
1️⃣ 一句话总结
这篇论文提出了一种通过测量大语言模型API的推理时间来估算其能耗的方法,帮助用户了解模型使用背后的能源成本,尤其适用于无法直接获取能耗数据的黑盒API服务。
The energy consumption of Large Language Models (LLMs) is raising growing concerns due to their adverse effects on environmental stability and resource use. Yet, these energy costs remain largely opaque to users, especially when models are accessed through an API -- a black box in which all information depends on what providers choose to disclose. In this work, we investigate inference time measurements as a proxy to approximate the associated energy costs of API-based LLMs. We ground our approach by comparing our estimations with actual energy measurements from locally hosted equivalents. Our results show that time measurements allow us to infer GPU models for API-based LLMs, grounding our energy cost estimations. Our work aims to create means for understanding the associated energy costs of API-based LLMs, especially for end users.
这太耗时了——探究将时间作为大语言模型能耗的代理指标 / This Is Taking Too Long -- Investigating Time as a Proxy for Energy Consumption of LLMs
这篇论文提出了一种通过测量大语言模型API的推理时间来估算其能耗的方法,帮助用户了解模型使用背后的能源成本,尤其适用于无法直接获取能耗数据的黑盒API服务。
源自 arXiv: 2603.15699