TU Darmstadt / ULB / TUbiblio

Method for the application of deep reinforcement learning for optimised control of industrial energy supply systems by the example of a central cooling system

Weigold, Matthias ; Ranzau, Heiko ; Schaumann, Sarah ; Kohne, Thomas ; Panten, Niklas ; Abele, Eberhard (2021):
Method for the application of deep reinforcement learning for optimised control of industrial energy supply systems by the example of a central cooling system.
In: CIRP Annals, 70 (1), pp. 17-20. Elsevier, ISSN 0007-8506,
DOI: 10.1016/j.cirp.2021.03.021,
[Article]

Abstract

This paper presents a method for data- and model-driven control optimisation for industrial energy supply systems (IESS) by means of deep reinforcement learning (DRL). The method consists of five steps, including system boundary definition and data accumulation, system modelling and validation, implementation of DRL algorithms, performance comparison and adaptation or application of the control strategy. The method is successfully applied to a simulation of an industrial cooling system using the PPO (proximal policy optimisation) algorithm. Significant reductions in electricity cost by 3% to 17% as well as reductions in CO2 emissions by 2% to 11% are achieved. The DRL-based control strategy is interpreted and three main reasons for the performance increase are identified. The DRL controller reduces energy cost by utilizing the storage capacity of the cooling system and moving electricity demand to times of lower prices. Additionally, the DRL-based control strategy for cooling towers as well as compression chillers reduces electricity cost and wear-related cost alike.

Item Type: Article
Erschienen: 2021
Creators: Weigold, Matthias ; Ranzau, Heiko ; Schaumann, Sarah ; Kohne, Thomas ; Panten, Niklas ; Abele, Eberhard
Title: Method for the application of deep reinforcement learning for optimised control of industrial energy supply systems by the example of a central cooling system
Language: English
Abstract:

This paper presents a method for data- and model-driven control optimisation for industrial energy supply systems (IESS) by means of deep reinforcement learning (DRL). The method consists of five steps, including system boundary definition and data accumulation, system modelling and validation, implementation of DRL algorithms, performance comparison and adaptation or application of the control strategy. The method is successfully applied to a simulation of an industrial cooling system using the PPO (proximal policy optimisation) algorithm. Significant reductions in electricity cost by 3% to 17% as well as reductions in CO2 emissions by 2% to 11% are achieved. The DRL-based control strategy is interpreted and three main reasons for the performance increase are identified. The DRL controller reduces energy cost by utilizing the storage capacity of the cooling system and moving electricity demand to times of lower prices. Additionally, the DRL-based control strategy for cooling towers as well as compression chillers reduces electricity cost and wear-related cost alike.

Journal or Publication Title: CIRP Annals
Journal volume: 70
Number: 1
Publisher: Elsevier
Uncontrolled Keywords: CO2 reduced production, Energy Efficiency, Machine learning
Divisions: 16 Department of Mechanical Engineering
16 Department of Mechanical Engineering > Institute of Production Technology and Machine Tools (PTW)
Date Deposited: 10 Nov 2021 07:18
DOI: 10.1016/j.cirp.2021.03.021
Official URL: https://www.sciencedirect.com/science/article/abs/pii/S00078...
Export:
Suche nach Titel in: TUfind oder in Google
Send an inquiry Send an inquiry

Options (only for editors)
Show editorial Details Show editorial Details