Reinforcement Learning-Based BEMS Architecture for Energy Usage Optimization
Abstract
:1. Introduction
1.1. AI-Based Energy Management
1.2. Intelligent Energy Data Analysis
1.3. Purpose-Oriented Energy Saving and Optimization
1.4. The Purpose of This Study
- Main scope
- Smart IoT-based cost-effective BEMS for existing buildings
- RL-based energy management model
- ML-based data analysis for time-series data reasoning
- Purpose-oriented energy saving and optimization methodology based on RL
- Total procedure
- Step 1: Collecting data by installing a smart IoT device inside the building (Section 4).
- Step 2: Establishing the relationship between the user and the building based on RL (Section 3).
- Step 3: Inferring the collected data based on ML (Section 5).
- Step 4: RL-based building energy optimization by establishing an HVAC schedule to be applied in the future based on the inferred data (Section 6).
2. Related Works
2.1. ANN (Artificial Neural Network)-Based BEMS Modeling
2.2. Reinforcement Learning (RL)-Based BEMS Modeling
2.3. Intelligent BEMS for Energy Optimization
2.4. Novel Energy Saving Routing Algorithm with Q-Learning Algorithm
2.5. Problems of Existing System
- Problem 1: Simple data prediction system using AI: Simple data analysis is only suitable for weather predictions, such as temperature and fine dust predictions.
- Problem 2: Energy savings via simple facility replacements for existing buildings: It is necessary to replace existing low-efficiency, old facilities to build BEMSs in existing buildings. However, these methods are expected to be expensive and difficult to install and require additional installation costs.
- Problem 3: Passive energy saving measures: There is a lack of precise goal setting for energy saving.
2.6. Merits of the Proposed System
- Solution 1: Reinforcement learning (RL)-based building energy optimization: This represents a more advanced AI-based method that functions via interactive exchanges between the RL-based user and HVAC.
- Solution 2: IoT-based lightweight and cost-effective system for existing buildings: Energy management is performed by installing a lightweight and cost-effective IoT system rather than replacing expensive equipment to build BEMSs in existing buildings.
- Solution 3: Purpose-oriented energy saving schedules: Energy-saving plans that are active, rather than passive, are proposed. Existing energy-saving methods are insufficient for achieving building energy savings up to 25%. Purpose-oriented energy saving plans means assuming that a 25% energy savings has already been achieved and then gradually complementing the user’s complaints by controlling the HVAC.
3. System Architecture
3.1. RL-Based System Architecture
- -
- User-BEMS side
- Action A (User request)
- Reward A (User Satisfaction or user dissatisfaction)
- -
- BEMS-HVAC system side
- Action B (Energy consumption)
- Reward B (Energy saving)
3.1.1. User-BEMS Side
3.1.2. BEMS-HVAC System Side
3.1.3. Optimization
4. System Configuration Test-Bed
4.1. System Installation
4.2. System Configuration
4.3. System Flow and Scenarios
5. Data Analysis
5.1. Temperature Data Analysis
5.2. Inflection Point Reasoning
5.3. Purpose-Oriented Optimization Method
6. Energy Usage Optimization
6.1. HVAC Scheduling Optimization for Energy Saving
6.2. RL-Based Algorithm
6.2.1. Markov Decision Process (MDP)
6.2.2. The Bellman Equation and Optimality
6.3. RL-Based Optimization
- It consumes unnecessary energy even though it already satisfies QoS
- Energy is saved too much and QoS is not satisfied
6.4. Implementation
6.4.1. Step 1: Initialize Random Value Function of A and B
6.4.2. Step 2: For Each State, Calculate Qa (s, a) and Qb (s, a): Q Table
6.4.3. Step 3: Update the Value Function with the Max Value of Qa(s, a) and Qb(s, a): Q Function Table
6.4.4. Step 4: Estimated Optimal Value Function Table
7. Future Prospects and Business Model (BM)
8. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Mataloto, B.; Ferreira, J.C.; Cruz, N. LoBEMS—IoT for Building and Energy Management Systems. Electronics 2019, 8, 763. [Google Scholar] [CrossRef] [Green Version]
- Levermore, G.J. Building Energy Management Systems: An Application to Heating and Control; E&FN Spon London: London, UK, 1992. [Google Scholar]
- Marinakis, V.; Doukas, H. An advanced IoT-based system for intelligent energy management in buildings. Sensors 2018, 18, 610. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Telecommunications Technology Association. Available online: http://www.tta.or.kr (accessed on 6 January 2020).
- Griego, D.; Krarti, M.; Hernandez-Guerrero, A. Energy efficiency optimization of new and existing office buildings in Guanajuato, Mexico. Sustain. Cities Soc. 2015, 17, 132–140. [Google Scholar] [CrossRef]
- Vikram, N.; Harish, K.; Nihaal, M.; Umesh, R.; Shetty, A.; Kumar, A. A low cost home automation system using Wi-Fi based wireless sensor network incorporating Internet of Things (IoT). In Proceedings of the 2017 IEEE 7th International Advance Computing Conference (IACC), Hyderabad, Germany, 5–7 January 2017; pp. 174–178. [Google Scholar]
- Peters, L.; Saidin, H. IT and the mass customization of services: The challenge of implementation. Int. J. Inf. Manag. 2000, 20, 103–119. [Google Scholar] [CrossRef]
- Park, S.; Park, S.; Byun, J.; Park, S. Design of a mass-customization-based cost-effective Internet of Things sensor system in smart building spaces. Int. J. Distrib. Sens. Netw. 2016, 12, 1550147716660895. [Google Scholar] [CrossRef] [Green Version]
- Li, S.; Xu, L.D.; Zhao, S. The internet of things: A survey. Inf. Syst. Front. 2015, 17, 243–259. [Google Scholar] [CrossRef]
- O’Dwyer, E.; Pan, I.; Acha, S.; Shah, N.J. Smart energy systems for sustainable smart cities: Current developments, trends and future directions. Appl. Energy 2019, 237, 581–597. [Google Scholar] [CrossRef] [Green Version]
- Qin, S.J.; Chiang, L.H. Advances and opportunities in machine learning for process data analytics. Comput. Chem. Eng. 2019, 126, 465–473. [Google Scholar] [CrossRef]
- Dhar, V. Data science and prediction. Commun. ACM 2013, 56, 64–73. [Google Scholar] [CrossRef]
- Leek, J. The key word in “data science” is not data, it is science. Available online: http://simplystatistics.org/2013/12/12/the-key-word-in-data-science-is-not-data-it-is-science/ (accessed on 6 January 2020).
- Hayashi, C. What is data science? Fundamental concepts and a heuristic example. In Data Science, Classification, and Related Methods; Springer: Tokyo, Japan, 1998; pp. 40–51. [Google Scholar]
- Wikipedia, Data Science. Available online: https://en.wikipedia.org/w/index.php?title=Data_science&oldid=934421711 (accessed on 6 January 2020).
- Ogudo, K.A.; Muwawa Jean Nestor, D.; Ibrahim Khalaf, O.; Daei Kasmaei, H. A device performance and data analytics concept for smartphones’ IoT services and machine-type communication in cellular networks. Symmetry 2019, 11, 593. [Google Scholar] [CrossRef] [Green Version]
- Zhao, L.; Wang, J.; Liu, J.; Kato, N. Routing for crowd management in smart cities: A deep reinforcement learning perspective. IEEE Commun. Mag. 2019, 57, 88–93. [Google Scholar] [CrossRef]
- Peng, W.; Gao, W.; Liu, J. AI-enabled massive devices multiple access for smart city. IEEE Internet Things J. 2019, 6, 7623–7634. [Google Scholar] [CrossRef]
- Mohammadi, M.; Al-Fuqaha, A.; Guizani, M.; Oh, J.-S. Semisupervised deep reinforcement learning in support of IoT and smart city services. IEEE Internet Things J. 2017, 5, 624–635. [Google Scholar] [CrossRef] [Green Version]
- He, Y.; Yu, F.R.; Zhao, N.; Leung, V.C.; Yin, H. Software-defined networks with mobile edge computing and caching for smart cities: A big data deep reinforcement learning approach. IEEE Commun. Mag. 2017, 55, 31–37. [Google Scholar] [CrossRef]
- Tom, R.J.; Sankaranarayanan, S.; Rodrigues, J.J. Smart Energy Management and Demand Reduction by Consumers and Utilities in an IoT-Fog-Based Power Distribution System. IEEE Internet Things J. 2019, 6, 7386–7394. [Google Scholar] [CrossRef]
- Zhao, H.-X.; Magoulès, F. A review on the prediction of building energy consumption. Renew. Sustain. Energy Rev. 2012, 16, 3586–3592. [Google Scholar] [CrossRef]
- Kalogirou, S.A. Applications of artificial neural-networks for energy systems. Appl. Energy 2000, 67, 17–35. [Google Scholar] [CrossRef]
- Chae, Y.T.; Horesh, R.; Hwang, Y.; Lee, Y.M. Artificial neural network model for forecasting sub-hourly electricity usage in commercial buildings. Energy Build. 2016, 111, 184–194. [Google Scholar] [CrossRef]
- Kalogirou, S.A.; Bojic, M. Artificial neural networks for the prediction of the energy consumption of a passive solar building. Energy 2000, 25, 479–491. [Google Scholar] [CrossRef]
- Liu, S.; Henze, G.P. Experimental analysis of simulated reinforcement learning control for active and passive building thermal storage inventory: Part 1. Theoretical foundation. Energy Build. 2006, 38, 142–147. [Google Scholar] [CrossRef]
- Liu, S.; Henze, G.P. Experimental analysis of simulated reinforcement learning control for active and passive building thermal storage inventory: Part 2: Results and analysis. Energy Build. 2006, 38, 148–161. [Google Scholar] [CrossRef]
- Yu, Z.; Dexter, A. Online tuning of a supervisory fuzzy controller for low-energy building system using reinforcement learning. Control Eng. Pract. 2010, 18, 532–539. [Google Scholar] [CrossRef]
- Kantamneni, A.; Brown, L.E.; Parker, G.; Weaver, W.W. Survey of multi-agent systems for microgrid control. Eng. Appl. Artif. Intell. 2015, 45, 192–203. [Google Scholar] [CrossRef]
- Dimeas, A.; Hatziargyriou, N. Multi-agent reinforcement learning for microgrids. In Proceedings of the IEEE PES General Meeting, Providence, RI, USA, 25–29 July 2010; pp. 1–8. [Google Scholar]
- Vázquez-Canteli, J.R.; Ulyanin, S.; Kämpf, J.; Nagy, Z. Fusing TensorFlow with building energy simulation for intelligent energy management in smart cities. Sustain. Cities Soc. 2019, 45, 243–257. [Google Scholar] [CrossRef]
- Brandi, S.; Piscitelli, M.S.; Martellacci, M.; Capozzoli, A. Deep Reinforcement Learning to optimise indoor temperature control and heating energy consumption in buildings. Energy Build. 2020, 110225. [Google Scholar] [CrossRef]
- Wang, F.; Zhou, L.; Ren, H.; Liu, X.; Talari, S.; Shafie-khah, M.; Catalão, J.P. Multi-objective optimization model of source–load–storage synergetic dispatch for a building energy management system based on TOU price demand response. IEEE Trans. Ind. Appl. 2017, 54, 1017–1028. [Google Scholar] [CrossRef]
- Wang, W.; Hong, T.; Li, N.; Wang, R.Q.; Chen, J. Linking energy-cyber-physical systems with occupancy prediction and interpretation through WiFi probe-based ensemble classification. Appl. Energy 2019, 236, 55–69. [Google Scholar] [CrossRef] [Green Version]
- Degha, H.E.; Laallam, F.Z.; Said, B. Intelligent context-awareness system for energy efficiency in smart building based on ontology. Sustain. Comput. Inform. Syst. 2019, 21, 212–233. [Google Scholar] [CrossRef]
- Park, H. Human Comfort-Based-Home Energy Management for Demand Response Participation. Energies 2020, 13, 2463. [Google Scholar] [CrossRef]
- Dey, M.; Rana, S.P.; Dudley, S. A Case Study Based Approach for Remote Fault Detection Using Multi-Level Machine Learning in A Smart Building. Smart Cities 2020, 3, 21. [Google Scholar] [CrossRef]
- Dey, M.; Rana, S.P.; Dudley, S. Smart building creation in large scale HVAC environments through automated fault detection and diagnosis. Future Gener. Comput. Syst. 2020, 108, 950–966. [Google Scholar] [CrossRef]
- Yang, C.; Gunay, B.; Shi, Z.; Shen, W. Machine Learning-Based Prognostics for Central Heating and Cooling Plant Equipment Health Monitoring. IEEE Trans. Autom. Sci. Eng. 2020. [Google Scholar] [CrossRef]
- Jafarinejad, T.; Erfani, A.; Fathi, A.; Shafii, M.B. Bi-level energy-efficient occupancy profile optimization integrated with demand-driven control strategy: University building energy saving. Sustain. Cities Soc. 2019, 48, 101539. [Google Scholar] [CrossRef]
- Zhang, D.; Zhang, T.; Liu, X. Novel self-adaptive routing service algorithm for application in VANET. Appl. Intell. 2019, 49, 1866–1879. [Google Scholar] [CrossRef]
- Zhang, D.; Li, G.; Zheng, K.; Ming, X.; Pan, Z.-H. An energy-balanced routing method based on forward-aware factor for wireless sensor networks. IEEE Trans. Ind. Inform. 2013, 10, 766–773. [Google Scholar] [CrossRef]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Ravichandiran, S. Hands-on Reinforcement Learning with Python: Master Reinforcement and Deep Reinforcement Learning Using OpenAI Gym and TensorFlow; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
- Zeiler, W.; Aduda, K.; Thomassen, T. Integral BEMS controlled LVPP in the smart grid: An approach to a complex distributed system of the built environment for sustainability. Int. J. Design Sci. Technol. 2016, 22, 81–108. [Google Scholar]
- Willmott, C.J.; Matsuura, K. Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance. Clim. Res. 2005, 30, 79–82. [Google Scholar] [CrossRef]
- Ravichandiran, S.; Saito, S.; Shanmugamani, R.; Wenzhuo, Y. Python Reinforcement Learning: Solve Complex Real-World Problems by Mastering Reinforcement Learning Algorithms Using OpenAI Gym and TensorFlow; Packt Publishing: Birmingham, UK, 2019. [Google Scholar]
- Zhu, Y.; Halpern, M.; Reddi, V.J. Event-based scheduling for energy-efficient qos (eqos) in mobile web applications. In Proceedings of the 2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA), Burlingame, CA, USA, 7–11 February 2015; pp. 137–149. [Google Scholar]
- Saito, S.; Wenzhuo, Y.; Shanmugamani, R. Python Reinforcement Learning Projects: Eight Hands-on Projects Exploring Reinforcement Learning Algorithms Using TensorFlow; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
- Lapan, M. Deep Reinforcement Learning Hands-On: Apply Modern RL Methods, with Deep Q-Networks, Value Iteration, Policy Gradients, TRPO, AlphaGo Zero and More; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
Existing System | Proposed System |
---|---|
|
|
Classification | A | B |
---|---|---|
Agent |
|
|
Environment |
|
|
Action |
|
|
State |
|
|
Reward |
|
|
Final Purpose |
|
Building Appearance | Characteristics |
---|---|
Building structure
| |
Smart-IoT Installation Site | |
26 Floor—Total 4 Rooms | 25 Floor—Total 4 Rooms |
15 Floor—Total 8 Room | 3 Floor—Total 8 Room |
Items | Characteristics | Uses | |
---|---|---|---|
|
|
| |
|
|
| |
|
|
| |
|
|
|
Rapidly | Slowly | |
---|---|---|
Rising | Sunlight shines into the room | HVAC system on |
Descending | Ventilation due to windows | HVAC system off |
Inflection Point: + | Inflection Point: − | |
---|---|---|
Meaning |
|
|
Reasoning 1 |
|
|
Reasoning 2 |
|
|
Reasoning 3 |
|
|
HVAC System On | HVAC System Off | |
---|---|---|
Case 1: Temperature Data |
|
|
Case 2: Power Data (Fan Coil) |
|
|
Classification | Operation Time | Time Zone | Expected Energy Saving Rate |
---|---|---|---|
Current System | 8 | Unassigned | - |
Proposed System | 6 | AM 3 h, PM 3 h | 20–25% |
Element | Description |
---|---|
S | A set of states (s) the agent can actually be in. |
A | A set of actions (a) that can be performed by an agent for moving from one state to another. |
Discount factor, which controls the importance of immediate and future rewards. | |
Transition probability, which is the probability of moving from one state (s) to another (s’) by performing some action a. | |
Reward probability, which is the probability of an agent acquiring a reward for moving from one state (s) to another (s’) by performing some action a. |
Classification | A | B |
---|---|---|
Agent |
|
|
Environment |
|
|
Action |
|
|
State |
|
|
Reward |
|
|
Classification | Inflection Point: + | Inflection Point: − |
---|---|---|
Reward A |
|
|
Reward B |
|
|
State A (Temperature) | Value A | State B (Energy Saving) | Value B |
---|---|---|---|
T1 (21 °C) | 0 | P1 (21%) | 0 |
T2 (22 °C) | 0 | P2 (22%) | 0 |
T3 (23 °C) | 0 | P3 (23%) | 0 |
T4 (24 °C) | 0 | P4 (24%) | 0 |
T5 (25 °C) | 0 | P5 (25%) | 0 |
State | Action | Next State | State | Action | Next State | ||||
---|---|---|---|---|---|---|---|---|---|
T1 | + | T1 | 0.3 | 1 | P1 | 1 | P1 | 0.2 | 0.8 |
T1 | + | T2 | 0.4 | 1 | P1 | 1 | P2 | 0.5 | 0.6 |
T1 | + | T3 | 0.2 | 1 | P1 | 1 | P3 | 0.4 | 0.4 |
T1 | + | T4 | 0.1 | 0.6 | P1 | 1 | P4 | 0.3 | 0.2 |
T1 | + | T5 | 0 | 0.7 | P1 | 1 | P5 | 0.2 | 0.1 |
T1 | - | T1 | 0.3 | 0.2 | P1 | 0 | P1 | 0.3 | 0.1 |
T1 | - | T2 | 0 | −0.2 | P1 | 0 | P2 | 0.2 | −0.2 |
T1 | - | T3 | 0 | −0.2 | P1 | 0 | P3 | 0.1 | −0.4 |
T1 | - | T4 | 0 | −0.2 | P1 | 0 | P4 | 0 | −0.6 |
T1 | - | T5 | 0 | −0.3 | P1 | 0 | P5 | 0 | −0.8 |
T2 | + | T1 | 0.1 | −0.1 | P2 | 1 | P1 | 0.1 | 1 |
T2 | + | T2 | 0.3 | 0.2 | P2 | 1 | P2 | 0.2 | 0.8 |
T2 | + | T3 | 0.4 | 1 | P2 | 1 | P3 | 0.5 | 0.6 |
T2 | + | T4 | 0.2 | 0.5 | P2 | 1 | P4 | 0.4 | 0.4 |
T2 | + | T5 | 0.1 | 0.6 | P2 | 1 | P5 | 0.3 | 0.2 |
T2 | - | T1 | 0.4 | 1 | P2 | 0 | P1 | 0.5 | 0.3 |
T2 | - | T2 | 0.3 | 0.2 | P2 | 0 | P2 | 0.3 | 0.1 |
T2 | - | T3 | 0.1 | −0.2 | P2 | 0 | P3 | 0.2 | −0.2 |
T2 | - | T4 | 0 | −0.2 | P2 | 0 | P4 | 0.1 | −0.4 |
T2 | - | T5 | 0 | −0.2 | P2 | 0 | P5 | 0 | −0.6 |
T3 | + | T1 | 0 | −0.2 | P3 | 1 | P1 | 0 | 1 |
T3 | + | T2 | 0 | −0.1 | P3 | 1 | P2 | 0.1 | 1 |
T3 | + | T3 | 0.2 | 0.2 | P3 | 1 | P3 | 0.2 | 0.8 |
T3 | + | T4 | 0.3 | 1 | P 3 | 1 | P4 | 0.5 | 0.6 |
T3 | + | T5 | 0.2 | 1 | P 3 | 1 | P5 | 0.4 | 0.4 |
T3 | - | T1 | 0.3 | 0.4 | P3 | 0 | P1 | 0.4 | 0.5 |
T3 | - | T2 | 0.4 | 1 | P3 | 0 | P2 | 0.5 | 0.3 |
T3 | - | T3 | 0.2 | 1 | P3 | 0 | P3 | 0.3 | 0.1 |
T3 | - | T4 | 0 | −0.2 | P3 | 0 | P4 | 0.2 | −0.2 |
T3 | - | T5 | 0 | −0.2 | P3 | 0 | P5 | 0.1 | −0.4 |
T4 | + | T1 | 0 | −0.3 | P4 | 1 | P1 | 0 | 1 |
T4 | + | T2 | 0 | −0.2 | P4 | 1 | P2 | 0 | 1 |
T4 | + | T3 | 0 | −0.1 | P4 | 1 | P3 | 0.1 | 1 |
T4 | + | T4 | 0.2 | 1 | P4 | 1 | P4 | 0.2 | 0.8 |
T4 | + | T5 | 0.3 | 0.4 | P4 | 1 | P5 | 0.5 | 0.6 |
T4 | - | T1 | 0.2 | 0.5 | P4 | 0 | P1 | 0.3 | 0.7 |
T4 | - | T2 | 0.3 | 0.4 | P4 | 0 | P2 | 0.4 | 0.5 |
T4 | - | T3 | 0.4 | 1 | P4 | 0 | P3 | 0.5 | 0.3 |
T4 | - | T4 | 0.2 | 1 | P4 | 0 | P4 | 0.3 | 0.1 |
T4 | - | T5 | 0.1 | −0.2 | P4 | 0 | P5 | 0.2 | −0.2 |
T5 | + | T1 | 0 | −0.4 | P5 | 1 | P1 | 0 | 1 |
T5 | + | T2 | 0 | −0.3 | P5 | 1 | P2 | 0 | 1 |
T5 | + | T3 | 0 | −0.2 | P5 | 1 | P3 | 0 | 1 |
T5 | + | T4 | 0 | −0.1 | P5 | 1 | P4 | 0.1 | 1 |
T5 | + | T5 | 0.2 | 1 | P5 | 1 | P5 | 0.2 | 0.8 |
T5 | - | T1 | 0.1 | 0.6 | P5 | 0 | P1 | 0.2 | 0.9 |
T5 | - | T2 | 0.2 | 0.5 | P5 | 0 | P2 | 0.3 | 0.7 |
T5 | - | T3 | 0.3 | 0.4 | P5 | 0 | P3 | 0.4 | 0.5 |
T5 | - | T4 | 0.4 | 0.3 | P5 | 0 | P4 | 0.5 | 0.3 |
T5 | - | T5 | 0.2 | 1 | P5 | 0 | P5 | 0.3 | 0.1 |
State A | Action A | Value A | State B | Action B | Value B |
---|---|---|---|---|---|
T1 | + | 0.96 | P1 | 1 | 0.7 |
T1 | − | 0.06 | P1 | 0 | −0.05 |
T2 | + | 0.61 | P2 | 1 | 0.78 |
T2 | − | 0.61 | P2 | 0 | 0.1 |
T3 | + | 0.54 | P3 | 1 | 0.72 |
T3 | − | 0.72 | P3 | 0 | 0.3 |
T4 | + | 0.32 | P4 | 1 | 0.56 |
T4 | − | 0.8 | P4 | 0 | 0.55 |
T5 | + | 0.2 | P5 | 1 | 0.26 |
T5 | − | 0.6 | P5 | 0 | 0.77 |
State A | Action A | Value A | State B | Action B | Value B |
---|---|---|---|---|---|
T1 | + | 1.4136 | P1 | 1 | 1.384 |
T1 | − | 0.2328 | P1 | 0 | 0.2128 |
T2 | + | 1.0822 | P2 | 1 | 1.4046 |
T2 | − | 0.8234 | P2 | 0 | 0.5704 |
T3 | + | 0.8424 | P3 | 1 | 1.206 |
T3 | − | 1.1256 | P3 | 0 | 0.945 |
T4 | + | 0.524 | P4 | 1 | 0.9014 |
T4 | − | 1.3298 | P4 | 0 | 1.2724 |
T5 | + | 0.272 | P5 | 1 | 0.386 |
T5 | − | 1.0044 | P5 | 0 | 1.3238 |
State A | Action A | Value A | State B | Action B | Value B |
---|---|---|---|---|---|
T1 | + | 1.689036 | P1 | 1 | 1.964788 |
T1 | − | 0.314448 | P1 | 0 | 0.440032 |
T2 | + | 1.379596 | P2 | 1 | 1.937052 |
T2 | − | 1.041596 | P2 | 0 | 0.989092 |
T3 | + | 1.034964 | P3 | 1 | 1.648428 |
T3 | − | 1.369248 | P3 | 0 | 1.502736 |
T4 | + | 0.660368 | P4 | 1 | 1.182188 |
T4 | − | 1.654412 | P4 | 0 | 1.885912 |
T5 | + | 0.320528 | P5 | 1 | 0.4952 |
T5 | − | 1.336968 | P5 | 0 | 1.948352 |
State A | Action A | Value A | State B | Action B | Value B |
---|---|---|---|---|---|
T1 | + | 2.140461707 | P1 | 1 | 10.83261728 |
T1 | − | 0.445283107 | P1 | 0 | 3.714355894 |
T2 | + | 1.879474769 | P2 | 1 | 10.35792026 |
T2 | − | 1.398131522 | P2 | 0 | 7.037297157 |
T3 | + | 1.376917547 | P3 | 1 | 8.644272198 |
T3 | − | 1.768587559 | P3 | 0 | 9.777649313 |
T4 | + | 0.927345189 | P4 | 1 | 5.946635124 |
T4 | − | 2.197902471 | P4 | 0 | 11.29521464 |
T5 | + | 0.429064595 | P5 | 1 | 2.315948923 |
T5 | − | 1.908871623 | P5 | 0 | 11.55501165 |
State A | Value A | State B | Value B |
---|---|---|---|
T1 (21 °C) | 2.1404617 | P1 (21%) | 10.832617 |
T2 (22 °C) | 1.8794748 | P2 (22%) | 10.35792 |
T3 (23 °C) | 1.7685876 | P3 (23%) | 9.7776493 |
T4 (24 °C) | 2.1979025 | P4 (24%) | 11.295215 |
T5 (25 °C) | 1.9088716 | P5 (25%) | 11.555012 |
Classification | Details | Target | Advantages | |
---|---|---|---|---|
Business Model 1 |
|
|
|
|
Business Model 2 |
|
|
|
|
Business Model 3 |
|
|
|
|
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Park, S.; Park, S.; Choi, M.-i.; Lee, S.; Lee, T.; Kim, S.; Cho, K.; Park, S. Reinforcement Learning-Based BEMS Architecture for Energy Usage Optimization. Sensors 2020, 20, 4918. https://doi.org/10.3390/s20174918
Park S, Park S, Choi M-i, Lee S, Lee T, Kim S, Cho K, Park S. Reinforcement Learning-Based BEMS Architecture for Energy Usage Optimization. Sensors. 2020; 20(17):4918. https://doi.org/10.3390/s20174918
Chicago/Turabian StylePark, Sanguk, Sangmin Park, Myeong-in Choi, Sanghoon Lee, Tacklim Lee, Seunghwan Kim, Keonhee Cho, and Sehyun Park. 2020. "Reinforcement Learning-Based BEMS Architecture for Energy Usage Optimization" Sensors 20, no. 17: 4918. https://doi.org/10.3390/s20174918
APA StylePark, S., Park, S., Choi, M. -i., Lee, S., Lee, T., Kim, S., Cho, K., & Park, S. (2020). Reinforcement Learning-Based BEMS Architecture for Energy Usage Optimization. Sensors, 20(17), 4918. https://doi.org/10.3390/s20174918