skip to main content
10.1145/3061639.3062224acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article

Deep Reinforcement Learning for Building HVAC Control

Published:18 June 2017Publication History

ABSTRACT

Buildings account for nearly 40% of the total energy consumption in the United States, about half of which is used by the HVAC (heating, ventilation, and air conditioning) system. Intelligent scheduling of building HVAC systems has the potential to significantly reduce the energy cost. However, the traditional rule-based and model-based strategies are often inefficient in practice, due to the complexity in building thermal dynamics and heterogeneous environment disturbances. In this work, we develop a data-driven approach that leverages the deep reinforcement learning (DRL) technique, to intelligently learn the effective strategy for operating the building HVAC systems. We evaluate the performance of our DRL algorithm through simulations using the widely-adopted EnergyPlus tool. Experiments demonstrate that our DRL-based algorithm is more effective in energy cost reduction compared with the traditional rule-based approach, while maintaining the room temperature within desired range.

References

  1. E. Barrett and S. Linder. Autonomous HVAC Control, A Reinforcement Learning Approach. Springer, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  2. L. Bottou. Large-scale machine learning with stochastic gradient descent. Proceedings of COMPSTAT. 2010.Google ScholarGoogle ScholarCross RefCross Ref
  3. G. T. Costanzo and et al. Experimental analysis of data-driven control for a building heating system. CoRR, abs/1507.03638, 2015.Google ScholarGoogle Scholar
  4. EnergyPlus. https://energyplus.net/.Google ScholarGoogle Scholar
  5. D. Ernst and et al. Tree-based batch mode reinforcement learning. Journal of Machine Learning Research, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. P. Fazenda and et al. Using reinforcement learning to optimize occupant comfort and energy usage in hvac systems. Journal of Ambient Intelligence and Smart Environments, pages 675--690, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. K. He and et al. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. IEEE International Conference on Computer Vision, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. G. Hinton, N. Srivastava, and K. Swersky. Lecture 6a overview of mini---batch gradient descent. http://www.es.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf.Google ScholarGoogle Scholar
  9. B. Li and L. Xia. A multi-grid reinforcement learning method for energy conservation and comfort of HVAC in buildings. pages 444--449, 2015.Google ScholarGoogle Scholar
  10. Y. Ma and et al. Model predictive control for the operation of building cooling systems. IEEE Transactions on Control Systems Technology, 20(3):796--803, 2012.Google ScholarGoogle ScholarCross RefCross Ref
  11. M. Maasoumy and et al. Model-based hierarchical optimal control design for HVAC systems. DSCC, 2011.Google ScholarGoogle Scholar
  12. V. Mnih and et al. Human-level control through deep reinforcement learning. Nature 518.7540, 2015.Google ScholarGoogle Scholar
  13. National Solar Radiation Data Base. http://rredc.nrel.gov.Google ScholarGoogle Scholar
  14. D. Nikovski, J. Xu, and M. Nonaka. A method for computing optimal set-point schedules for HVAC systems. REHVA World Congress CLIMA, 2013.Google ScholarGoogle Scholar
  15. F. Oldewurtel and et al. Energy efficient building climate control using stochastic model predictive control and weather predictions. ACC, 2010.Google ScholarGoogle Scholar
  16. S. J. Olivieri and et al. Evaluation of commercial building demand response potential using optimal short-term curtailment of heating, ventilation, and air-conditioning loads. Journal of Building Performance Simulation, 2014.Google ScholarGoogle Scholar
  17. D. Ormoneit and S. Sen. Kernel-based reinforcement learning. Machine Learning, 49(2):161--178, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. M. Riedmiller. Neural Fitted Q Iteration -- First Experiences with a Data Efficient Neural Reinforcement Learning Method. Springer, 2005.Google ScholarGoogle Scholar
  19. D. Silver and et al. Mastering the game of go with deep neural networks and tree search. Nature, 529(7587), 2016.Google ScholarGoogle Scholar
  20. SCE. https://www.sce.com/NR/sc3/tm2/pdf/CE281.pdf.Google ScholarGoogle Scholar
  21. A. Standard. Standard 55-2004-thermal environmental conditions for human occupancy. ASHRAE Inc., 2004.Google ScholarGoogle Scholar
  22. D. Urieli and P. Stone. A learning agent for heat-pump thermostat control. AAMAS, 2013. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. U.S. DoE. Buildings energy data book.Google ScholarGoogle Scholar
  24. C. J. Watkins and P. Dayan. Q-learning. Machine learning, 8(3-4):279--292, 1992. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. T. Wei, Q. Zhu, and M. Maasoumy. Co-scheduling of HVAC control, EV charging and battery usage for building energy efficiency. ICCAD, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. M. Wetter. Co-simulation of building energy and control systems with the building controls virtual test bed. Journal of Building Performance Simulation, 2011.Google ScholarGoogle ScholarCross RefCross Ref
  27. L. Yang and et al. Reinforcement learning for optimal control of low exergy buildings. Applied Energy, 2015.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    DAC '17: Proceedings of the 54th Annual Design Automation Conference 2017
    June 2017
    533 pages
    ISBN:9781450349277
    DOI:10.1145/3061639

    Copyright © 2017 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 18 June 2017

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited

    Acceptance Rates

    Overall Acceptance Rate1,770of5,499submissions,32%

    Upcoming Conference

    DAC '24
    61st ACM/IEEE Design Automation Conference
    June 23 - 27, 2024
    San Francisco , CA , USA

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader