Towards Accurate and Reliable Energy Measurement of NLP Models
October 11, 2020 ยท Entered Twilight ยท ๐ SUSTAINLP
"Last commit was 5.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: .gitignore, README.md, compute_efficiency_info.py, compute_wattsup_energy.py, extract_duration.py, extract_runtime_stats.py, plot.py, random_select_questions.py, requirements.txt, run_energy_squad.py, scripts.sh, wattsup.py
Authors
Qingqing Cao, Aruna Balasubramanian, Niranjan Balasubramanian
arXiv ID
2010.05248
Category
cs.CL: Computation & Language
Citations
38
Venue
SUSTAINLP
Repository
https://github.com/csarron/sustainlp2020-energy
โญ 5
Last Checked
1 month ago
Abstract
Accurate and reliable measurement of energy consumption is critical for making well-informed design choices when choosing and training large scale NLP models. In this work, we show that existing software-based energy measurements are not accurate because they do not take into account hardware differences and how resource utilization affects energy consumption. We conduct energy measurement experiments with four different models for a question answering task. We quantify the error of existing software-based energy measurements by using a hardware power meter that provides highly accurate energy measurements. Our key takeaway is the need for a more accurate energy estimation model that takes into account hardware variabilities and the non-linear relationship between resource utilization and energy consumption. We release the code and data at https://github.com/csarron/sustainlp2020-energy.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted