Reinforcement Learning-Based Energy Management of Smart Home with Rooftop Solar Photovoltaic System, Energy Storage System, and Home Appliances
Abstract
:1. Introduction
- We present an RL-based HEMS model that manages the optimal energy consumption of a smart home with a rooftop PV system, ESS, and smart home appliances. In the HEMS model, the Q-learning method is applied to the energy consumption scheduling of different home appliances (air conditioner, washing machine, and ESS), whereby the agent of each appliance determines the optimal policy independently to reduce its own electric cost within the consumer comfort level and the appliance operation characteristics. Furthermore, we propose an ANN model to learn the relationship between the indoor temperature and energy consumption of the air conditioner more accurately, which is integrated into the Q-learning module to achieve improved performance of the air conditioner agent.
- The simulation results confirm that the proposed RL method with the ANN can successfully reduce both the consumer electricity bill and dissatisfaction cost (for example, the indoor temperature and operating time interval of the washing machine within the consumer comfort settings). Moreover, we compare the performance of the proposed RL-based HEMS algorithm to that of the conventional mixed-integer linear programming (MILP)-based HEMS algorithm, and verify that the proposed approach can achieve greater energy savings than the conventional approach under various penalty parameter settings in the reward function of the appliance agent.
2. Related Research
3. System Model for HEMS
3.1. Preliminary
- Controllable appliance (): A controllable appliance is an appliance of which the operation is scheduled and controlled by the HEMS. The operation characteristics categorize controllable appliances into reducible appliances () and shiftable appliances (). An example of a reducible appliance is an air conditioner, known as a thermostatically controllable load, in which the energy consumption can be curtailed to reduce the electricity bill. However, under the TOU pricing scheme, the energy consumption of a shiftable appliance can be shifted from one time slot to another to minimize the total electricity cost. A shiftable appliance has two load types: (1) a non-interruptible load (), and (2) an interruptible load (). The operation of shiftable appliances with non-interruptible loads must not be stopped by the HEMS control during the appliance task period. For example, a washing machine must perform a washing cycle prior to drying. A shiftable appliance with an interruptible load may be interrupted at any time. For example, the HEMS must terminate the discharging process and initiate the charging process of the ESS instantly when the PV power generation is greater than the load demand.
- Uncontrollable appliance (): An uncontrollable appliance, such as a TV, PC, or lighting, cannot be scheduled and operated by the HEMS. Therefore, maintains the fixed energy consumption scheduling.
3.2. Conventional HEMS Optimization Formulation
3.2.1. Objective Function
3.2.2. Net Power Consumption
3.2.3. Operating Characteristics for Controllable Appliances
4. Formulation of RL- and ANN-Based Home Energy Management
4.1. Home Energy Management via Q-Learning
4.1.1. State Space
4.1.2. Action Space
4.1.3. Reward
4.2. Prediction of Indoor Temperature via ANN
Algorithm 1: Q-learning-based energy management of smart home with PV system, ESS, and home appliances. |
1 Initialize each appliance’s energy demand, dissatisfaction parameters, and Q-learning parameters 2 %%Learning with ANN for temperature prediction of AC agent 3 Indoor temperature at time period → 4 Minimum and maximum value of consumer’s comfort temperature range → , 5 Predicted outdoor temperature at time t → 6 Energy consumption of AC agent at time t → 7 Predicted indoor temperature at time t → 8 Learning process with ANN and approximate the temperature prediction model 9 10 Initialize Q-value of each agent 19 Find optimal policy with largest Q-value |
5. Numerical Examples
5.1. Simulation Setup
5.2. Performance of the Proposed RL-Based HEMS Algorithm
5.3. Impact of Different Parameters in Reward Function on the Proposed Algorithm
5.4. Impact of ANN on AC Agent Performance
5.5. Performance Comparison between MILP- and RL-Based HEMS
6. Discussion
6.1. Wholesale and Retail Electricity Markets under Real-Time Pricing (RTP)
6.2. Electric Vehicle (EV) Integration
6.3. Constraint of the Lifetime for ESS
7. Conclusions
Author Contributions
Funding
Conflicts of Interest
Nomenclature
TOU price at time slot t | |
Penalty for consumer thermal discomfort cost | |
Energy consumption of appliance a at time slot t | |
Net energy consumption at time slot t | |
Maximum (Minimum) energy consumption of appliance a | |
Charging energy of ESS a at time slot t | |
Discharging energy of ESS a at time slot t | |
Maximum (Minimum) charging energy of ESS a | |
Maximum (Minimum) discharging energy of ESS a | |
State of energy of ESS a at time slot t | |
Maximum (Minimum) state of energy of ESS a | |
Charging efficiency of ESS a | |
Discharging efficiency of ESS a | |
Predicted PV generation energy at time slot t | |
Indoor temperature at time slot t | |
Predicted outdoor temperature at time slot t | |
Maximum (Minimum) preferred indoor temperature | |
Thermal characteristic for AC | |
Consumer preferred indoor temperature | |
Consumer preferred starting time for WM | |
Consumer preferred finishing time for WM | |
Binary charging and discharging state of ESS a at time slot t: “1” for charging, “0” for discharging | |
Binary consumption state of non-interruptible shiftable appliance a at time slot t: | |
“1” for consumption, “0” otherwise | |
State at time slot t | |
Action at time slot t | |
Reward of WM agent at time slot t | |
Reward of AC agent at time slot t | |
Reward of ESS agent at time slot t | |
Learning rate in RL | |
Discount factor in RL | |
Energy unit in discrete set of actions for AC agent | |
Energy unit in discrete set of actions for ESS | |
Penalty for early (late) operation of WM in reward function | |
Penalty for overcharging (undercharging) of ESS in reward function | |
Penalty for consumer thermal discomfort cost in reward function | |
Set of time slots | |
Set of appliances | |
Set of reducible appliances | |
Set of shiftable appliances with non-interruptible loads | |
Set of shiftable appliances with interruptible loads | |
Set of uncontrollable appliances | |
Set of states for WM | |
Set of states for AC | |
Set of states for ESS | |
Set of actions for WM | |
Set of actions for AC | |
Set of actions for ESS |
References
- Mosavi, A.; Salimi, M.; Ardabili, S.F.; Rabczuk, T.; Shamshirband, S.; Varkonyi-Koczy, A.R. State of the art of machine learning models in energy systems, a systematic Review. Energies 2019, 12, 1301. [Google Scholar] [CrossRef]
- Nejat, P.; Jomehzadeh, F.; Taheri, M.M.; Gohari, M.; Majid, M.Z.A. A global review of energy consumption, CO2 emissions and policy in the residential sector (with an overview of the top ten CO2 emitting countries). Renew. Sustain. Energy Rev. 2015, 43, 843–862. [Google Scholar] [CrossRef]
- Joo, I.-Y.; Choi, D.-H. Optimal household appliance scheduling considering consumer’s electricity bill target. IEEE Trans. Consum. Electron. 2017, 63, 19–27. [Google Scholar] [CrossRef]
- Nguyen, D.T.; Le, L.B. Joint optimization of electric vehicle and home energy scheduling considering user comfort preference. IEEE Trans. Smart Grid 2014, 5, 188–199. [Google Scholar] [CrossRef]
- Zhao, C.; Dong, S.; Li, F.; Song, Y. Optimal home energy management system with mixed types of loads. CSEE J. Power Energy 2014, 1, 29–37. [Google Scholar] [CrossRef]
- Althaher, S.; Mancarella, P.; Mutale, J. Automated Demand Response From Home Energy Management System Under Dynamic Pricing and Power and Comfort Constraints. IEEE Trans. Smart Grid 2015, 6, 1874–1883. [Google Scholar] [CrossRef]
- Tsui, K.M.; Chan, S.C. Demand response optimization for smart home scheduling under real-time pricing. IEEE Trans. Smart Grid 2012, 3, 1812–1821. [Google Scholar] [CrossRef]
- Paterakis, N.G.; Erdinc, O.; Pappi, I.N.; Bakirtzis, A.G.; Catalão, J.P.S. Coordinated operation of a neighborhood of smart households comprising electric vehicles, energy storage and distributed Generation. IEEE Trans. Smart Grid 2016, 7, 2736–2747. [Google Scholar] [CrossRef]
- Erdinc, O.; Paterakis, N.G.; Mendes, T.D.P.; Bakirtzis, A.G.; Catalão, J.P.S. Smart household operation considering bi-directional EV and ESS utilization by real-time pricing-based DR. IEEE Trans. Smart Grid 2015, 6, 1281–1291. [Google Scholar] [CrossRef]
- Liu, Y.; Xiao, L.; Yao, G. Pricing-based demand response for a smart home with various types of household appliances considering customer satisfaction. IEEE Access 2019. [Google Scholar] [CrossRef]
- Lee, S.; Kwon, B.; Lee, S. Joint energy management system of electric supply and demand in houses and buildings. IEEE Trans. Power Syst. 2014, 6, 2804–2812. [Google Scholar] [CrossRef]
- Luo, F.; Ranzi, G.; Wan, C.; Xu, Z.; Dong, Z.Y. A multistage home energy management system with residential photovoltaic penetration. IEEE Trans. Ind. Inf. 2019, 15, 116–126. [Google Scholar] [CrossRef]
- Ran, X.; Leng, S. Enhanced robust index model for load scheduling of a home energy local network with a load shifting strategy. IEEE Access 2019, 7, 19943–19953. [Google Scholar] [CrossRef]
- Wang, J.; Li, P.; Fang, K.; Zhou, Y. Robust optimization for household load scheduling with uncertain parameters. Appl. Sci. 2018, 8, 575. [Google Scholar] [CrossRef]
- Rastegar, M.; Fotuhi-Firuzabad, M.; Moeini-Aghtaie, M. Developing a two-level framework for residential energy management. IEEE Trans. Smart Grid 2018, 9, 1707–1717. [Google Scholar] [CrossRef]
- Joo, I.-Y.; Choi, D.-H. Distributed optimization framework for energy management of multiple smart homes with distributed energy resources. IEEE Access 2017, 5, 2169–3536. [Google Scholar] [CrossRef]
- Luo, F.; Kong, W.; Ranzi, G.; Dong, Z.Y. Optimal home energy management system with demand charge tariff and appliance operational dependencies. IEEE Trans. Smart Grid 2019. [Google Scholar] [CrossRef]
- Shareef, H.; Ahmed, M.S.; Mohamed, A.; Hassan, E.A. Review on home energy management system considering demand responses, smart technologies, and intelligent controllers. IEEE Access 2018, 6, 24498–24509. [Google Scholar] [CrossRef]
- Shaikh, P.H.; Nor, N.B.M.; Nallagownden, P.; Elamvazuthi, I.; Ibrahim, T. A review on optimized control systems for building energy and comfort management of smart sustainable buildings. Renew. Sustain. Energy Rev. 2014, 34, 409–429. [Google Scholar] [CrossRef]
- Li, Y.; Yan, Z.; Chen, S.; Xu, X.; Kang, C. Operation strategy of smart thermostats that self-learn user preferences. IEEE Trans. Smart Grid 2019. [Google Scholar] [CrossRef]
- Shi, H.; Xu, M.; Li, R. Deep learning for household load forecasting—A novel pooling deep RNN. IEEE Trans. Smart Grid 2018, 9, 5271–5280. [Google Scholar] [CrossRef]
- Zhang, D.; Li, S.; Sun, M.; O’Neill, Z. An optimal and learning-based demand response and home energy management system. IEEE Trans. Smart Grid 2016, 7, 1790–1801. [Google Scholar] [CrossRef]
- Liu, H.; Mi, X.; Li, Y. Comparison of two new intelligent wind speed forecasting approaches based on Wavelet Packet Decomposition, Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Artificial Neural Networks. Energy Convers. Manag. 2018, 155, 188–200. [Google Scholar] [CrossRef]
- Hao, Y.; Tian, C. The study and application of a novel hybrid system for air quality early-warning. Appl. Soft Comput. 2019, 74, 729–746. [Google Scholar] [CrossRef]
- Wang, D.; Luo, H.; Grunder, O.; Lin, Y.; Guo, H. Multi-step ahead electricity price forecasting using a hybrid model based on two-layer decomposition technique and BP neural network optimized by firefly algorithm. Appl. Energy 2017, 190, 390–407. [Google Scholar] [CrossRef]
- Magalhaes, S.M.C.; Leal, V.M.S.; Horta, I.M. Modelling the relationship between heating energy use and indoor temperatures in residential buildings through artificial neural networks considering occupant behavior. Energy Build. 2017, 151, 332–343. [Google Scholar] [CrossRef]
- Wei, T.; Wang, Y.; Zhu, Q. Deep reinforcement learning for building HVAC control. In Proceedings of the 2017 54th ACM/EDAC/IEEE Design Automation Conference (DAC), Austin, TX, USA, 18–22 June 2017; pp. 1–6. [Google Scholar] [CrossRef]
- Eller, L.; Siafara, L.C.; Sauter, T. Adaptive control for building energy management using reinforcement learning. In Proceedings of the 2018 IEEE International Conference on Industrial Technology (ICIT), Lyon, France, 20–22 February 2018; pp. 1–6. [Google Scholar] [CrossRef]
- Kim, S.; Lim, H. Reinforcement learning based energy management algorithm for smart energy buildings. Energies 2018, 11, 2010. [Google Scholar] [CrossRef]
- Zhou, S.; Hu, Z.; Gu, W.; Jiang, M.; Zhang, X.-P. Artificial intelligence based smart energy community management: A reinforcement learning approach. CSEE J. Power Energy 2019, 5, 840. [Google Scholar] [CrossRef]
- Lu, R.; Hong, S.H.; Yu, M. Demand response for home energy management using reinforcement learning and artificial neural network. IEEE Trans. Smart Grid 2019. [Google Scholar] [CrossRef]
- Mua, C.; Zhao, Q.; Sun, C.; Gao, Z. An ADDHP-based Q-learning algorithm for optimal tracking control of linear discrete-time systems with unknown dynamics. Appl. Soft Comput. 2019, 82, 1–13. [Google Scholar] [CrossRef]
- Mu, C.; Zhao, Q.; Gao, Z.; Sun, C. Q-learning solution for optimal consensus control of discrete-time multiagent systems using reinforcement learning. J. Frankl. Inst. 2019, 356, 6946–6967. [Google Scholar] [CrossRef]
- Si, J.; Harris, S.L.; Yfantis, E. A dynamic ReLU on neural network. In Proceedings of the 2018 IEEE 13th Dallas Circuits and Systems Conference (DCAS), Dallas, TX, USA, 12 November 2018; pp. 1–6. [Google Scholar] [CrossRef]
- Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Wu, T.; Alaywan, Z.; Papalexopoulos, A.D. Locational Marginal Price Calculations Using the Distributed-Slack Power-Flow Formulation. IEEE Trans. Power Syst. 2005, 20, 1188–1190. [Google Scholar] [CrossRef]
- Li, R.; Wu, Q.; Oren, S.S. Distribution Locational Marginal Pricing for Optimal Electric Vehicle Charging Management. IEEE Trans. Power Syst. 2014, 29, 203–211. [Google Scholar] [CrossRef]
- Mohsenzadeh, A.; Pang, C. Two stage residential energy management under distribution locational marginal pricing. Electr. Power Syst. Res. 2018, 154, 361–372. [Google Scholar] [CrossRef]
- Park, Y.-G.; Park, J.-B.; Kim, N.; Lee, K.Y. Linear Formulation for Short-Term Operational Scheduling of Energy Storage Systems in Power Grids. Energies 2017, 10, 207. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, S.; Choi, D.-H. Reinforcement Learning-Based Energy Management of Smart Home with Rooftop Solar Photovoltaic System, Energy Storage System, and Home Appliances. Sensors 2019, 19, 3937. https://doi.org/10.3390/s19183937
Lee S, Choi D-H. Reinforcement Learning-Based Energy Management of Smart Home with Rooftop Solar Photovoltaic System, Energy Storage System, and Home Appliances. Sensors. 2019; 19(18):3937. https://doi.org/10.3390/s19183937
Chicago/Turabian StyleLee, Sangyoon, and Dae-Hyun Choi. 2019. "Reinforcement Learning-Based Energy Management of Smart Home with Rooftop Solar Photovoltaic System, Energy Storage System, and Home Appliances" Sensors 19, no. 18: 3937. https://doi.org/10.3390/s19183937
APA StyleLee, S., & Choi, D.-H. (2019). Reinforcement Learning-Based Energy Management of Smart Home with Rooftop Solar Photovoltaic System, Energy Storage System, and Home Appliances. Sensors, 19(18), 3937. https://doi.org/10.3390/s19183937