Dynamic pricing and energy consumption scheduling with reinforcement learning

Byung Gook Kim, Yu Zhang, Mihaela Van Der Schaar, Jang Won Lee

Research output: Contribution to journalArticle

59 Citations (Scopus)

Abstract

In this paper, we study a dynamic pricing and energy consumption scheduling problem in the microgrid where the service provider acts as a broker between the utility company and customers by purchasing electric energy from the utility company and selling it to the customers. For the service provider, even though dynamic pricing is an efficient tool to manage the microgrid, the implementation of dynamic pricing is highly challenging due to the lack of the customer-side information and the various types of uncertainties in the microgrid. Similarly, the customers also face challenges in scheduling their energy consumption due to the uncertainty of the retail electricity price. In order to overcome the challenges of implementing dynamic pricing and energy consumption scheduling, we develop reinforcement learning algorithms that allow each of the service provider and the customers to learn its strategy without a priori information about the microgrid. Through numerical results, we show that the proposed reinforcement learning-based dynamic pricing algorithm can effectively work without a priori information about the system dynamics and the proposed energy consumption scheduling algorithm further reduces the system cost thanks to the learning capability of each customer.

Original languageEnglish
Article number7321806
Pages (from-to)2187-2198
Number of pages12
JournalIEEE Transactions on Smart Grid
Volume7
Issue number5
DOIs
Publication statusPublished - 2016 Sep

All Science Journal Classification (ASJC) codes

  • Computer Science(all)

Fingerprint Dive into the research topics of 'Dynamic pricing and energy consumption scheduling with reinforcement learning'. Together they form a unique fingerprint.

  • Cite this