Hostname: page-component-89b8bd64d-shngb Total loading time: 0 Render date: 2026-05-11T16:20:39.777Z Has data issue: false hasContentIssue false

Optimizing HVAC energy efficiency in low-energy buildings: a comparative analysis of reinforcement learning control strategies under Tehran climate conditions

Published online by Cambridge University Press:  22 August 2025

Mohammad Anvar Adibhesami
Affiliation:
School of Architecture and Environmental Design, Iran University of Science and Technology , Tehran, Iran
Amir Hassanzadeh*
Affiliation:
Department of Mechanical Engineering, Urmia University , Urmia, Iran
*
Corresponding author: Amir Hassanzadeh; Email: st_a.hassanzadeh@urmia.ac.ir

Abstract

This study investigates the incorporation of advanced heating, ventilation, and air conditioning (HVAC) systems with reinforcement learning (RL) control to enhance energy efficiency in low-energy buildings amid the extreme seasonal temperatures of Tehran. We conducted comprehensive simulation assessments using the EnergyPlus and HoneybeeGym platforms to evaluate two distinct reinforcement learning models: traditional Q-learning (Model A) and deep reinforcement learning (DRL) with neural networks (Model B). Model B consisted of a deep convolutional network architecture with 256 neurons in each hidden layer, employing rectified linear units as activation functions and the Adam optimizer at a learning rate of 0.001. The results demonstrated that the RL-managed systems resulted in a statistically significant reduction in energy-use intensity of 25 percent (p < 0.001), decreasing from 250 to 200 kWh/m² annually in comparison to the baseline scenario. The thermal comfort showed notable improvements, with the expected mean vote adjusting to 0.25, which falls within the ASHRAE Standard 55 comfort range, and the percentage of anticipated dissatisfaction reduced to 10%. Model B (DRL) demonstrated a 50 percent improvement in prediction accuracy over Model A, with a mean absolute error of 0.579366 compared to 1.140008 and a root mean square error of 0.689770 versus 1.408069. This indicates enhanced adaptability to consistent daily trends and irregular periodicities, such as weather patterns. The proposed reinforcement learning method achieved energy savings of 10–15 percent compared to both rule-based and model predictive control and approximately 10 percent improvement over rule-based control, while employing fewer building features than existing state-of-the-art control systems.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Figure 1. Conceptual integration framework for HVAC energy efficiency study.

Figure 1

Table 1. Literature review matrix

Figure 2

Figure 2. Tehran residential HVAC energy usage zones.

Figure 3

Table 2. Summary of collected HVAC energy usage data (Tehran)

Figure 4

Figure 3. Framework of simulation of this study.

Figure 5

Figure 4. Reinforcement learning implementation process.

Figure 6

Table 3. Simulation parameters

Figure 7

Figure 5. Methodology framework for results analysis.

Figure 8

Table 4. Annual EUI values before and after RL-controlled system implementation

Figure 9

Figure 6. Monthly EUI comparison.

Figure 10

Figure 7. Thermal comfort distribution.

Figure 11

Table 5. Thermal comfort metrics

Figure 12

Figure 8. Monthly operational cost comparison.

Figure 13

Table 6. Annual operational cost comparison

Figure 14

Table 7. Summary of reinforcement learning model parameters

Figure 15

Figure 9. Episode reward over time.

Figure 16

Figure 10. Actions selected by the RL agent. Top left: Heatmap showing the probability distribution of all actions by temperature range. Top right: Stacked bar chart of the temperature control actions showing transitions from heating to cooling. Bottom left: Dominant action selected for each temperature range.

Figure 17

Figure 11. Temperature humidity control performance. Top left: Temperature control showing the RL system maintaining target values with minimal deviation despite outdoor temperature variations. Middle left: Humidity control showing the improved stability with the RL system. Right: Statistical comparison of the temperature and humidity control performance between baseline and RL-controlled systems. Bottom: Detailed analysis of the system response to temperature setpoint change, demonstrating a faster response time and better stability with RL control.

Figure 18

Figure 12. Performance metrics visualization.

Figure 19

Table 8. Comprehensive performance metrics comparison

Figure 20

Figure 13. Sensitivity analysis of HVAC RL model parameters.

Figure 21

Table 9. Sensitivity analysis results

Submit a response

Comments

No Comments have been published for this article.