Large Foundation Models for Power Systems

December 12, 2023 ยท Entered Twilight ยท ๐Ÿ› IEEE Power & Energy Society General Meeting

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, Document_Summary, LICENSE, LLM_as_optimizer_gpt.py, LLM_optimizer_ev.py, Large_Foundation_Models_Power_Systems.pdf, README.md, Situation_Awareness, Wind_Power_Forecasting

Authors Chenghao Huang, Siyang Li, Ruohong Liu, Hao Wang, Yize Chen arXiv ID 2312.07044 Category eess.SY: Systems & Control (EE) Cross-listed cs.LG Citations 55 Venue IEEE Power & Energy Society General Meeting Repository https://github.com/chennnnnyize/LLM_PowerSystems โญ 72 Last Checked 1 month ago
Abstract
Foundation models, such as Large Language Models (LLMs), can respond to a wide range of format-free queries without any task-specific data collection or model training, creating various research and application opportunities for the modeling and operation of large-scale power systems. In this paper, we outline how such large foundation model such as GPT-4 are developed, and discuss how they can be leveraged in challenging power and energy system tasks. We first investigate the potential of existing foundation models by validating their performance on four representative tasks across power system domains, including the optimal power flow (OPF), electric vehicle (EV) scheduling, knowledge retrieval for power engineering technical reports, and situation awareness. Our results indicate strong capabilities of such foundation models on boosting the efficiency and reliability of power system operational pipelines. We also provide suggestions and projections on future deployment of foundation models in power system applications.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Systems & Control (EE)