Building operations contribute significantly to global carbon emissions, yet existing digital twins rely on static strategies that fail to adapt to dynamic conditions. This paper investigates whether an adaptive reinforcement learning framework with hierarchical transfer learning can optimize real-time energy use and comfort across heterogeneous buildings without extensive retraining. The paper develops and validates a proximal policy optimization controller integrated with a digital twin and 348 IoT sensors across three commercial buildings in a 14-month randomized crossover trial. The approach achieves 23.7 % and 14.9 % energy reductions compared to rule-based control and model predictive control, respectively, while maintaining 94.7 % comfort satisfaction and demonstrating 78 % transfer learning efficiency. These findings provide facility managers and grid operators with a scalable, hardware-validated approach that reduces operational costs and stabilizes demand response without compromising occupant comfort. Future work extends this hierarchical transfer framework to mixed-mode ventilation and district-level energy coordination to further enhance grid interactivity.
扫码关注我们
求助内容:
应助结果提醒方式:
