TY - JOUR
T1 - Policies for the dynamic traveling maintainer problem with alerts
AU - da Costa, Paulo
AU - Verleijsdonk, Peter
AU - Voorberg, Simon
AU - Akcay, Alp
AU - Kapodistria, Stella
AU - van Jaarsveld, Willem
AU - Zhang, Yingqian
PY - 2023/3/16
Y1 - 2023/3/16
N2 - Downtime of industrial assets such as wind turbines and medical imaging devices comes at a sharp cost. To avoid such downtime costs, companies seek to initiate maintenance just before failure. Unfortunately, this is challenging for the following two reasons: On the one hand, because asset failures are notoriously difficult to predict, even in the presence of real-time monitoring devices which signal early degradation. On the other hand, because the available resources to serve a network of geographically dispersed assets are typically limited. In this paper, we propose a novel model referred to as the dynamic traveling maintainer problem with alerts that incorporates these two challenges and we provide three solution approaches on how to dispatch the limited resources. Namely, we propose: (i) Greedy heuristic approaches that rank assets on urgency, proximity and economic risk; (ii) A novel traveling maintainer heuristic approach that optimizes short-term costs; and (iii) A deep reinforcement learning (DRL) approach that optimizes long-term costs. Each approach has different requirements concerning the available alert information. Experiments with small asset networks show that all methods can approximate the optimal policy when given access to complete condition information. For larger networks, the proposed methods yield competitive policies, with DRL consistently achieving the lowest costs.
AB - Downtime of industrial assets such as wind turbines and medical imaging devices comes at a sharp cost. To avoid such downtime costs, companies seek to initiate maintenance just before failure. Unfortunately, this is challenging for the following two reasons: On the one hand, because asset failures are notoriously difficult to predict, even in the presence of real-time monitoring devices which signal early degradation. On the other hand, because the available resources to serve a network of geographically dispersed assets are typically limited. In this paper, we propose a novel model referred to as the dynamic traveling maintainer problem with alerts that incorporates these two challenges and we provide three solution approaches on how to dispatch the limited resources. Namely, we propose: (i) Greedy heuristic approaches that rank assets on urgency, proximity and economic risk; (ii) A novel traveling maintainer heuristic approach that optimizes short-term costs; and (iii) A deep reinforcement learning (DRL) approach that optimizes long-term costs. Each approach has different requirements concerning the available alert information. Experiments with small asset networks show that all methods can approximate the optimal policy when given access to complete condition information. For larger networks, the proposed methods yield competitive policies, with DRL consistently achieving the lowest costs.
KW - Decision process
KW - Deep reinforcement learning
KW - Degradation process
KW - Maintenance
KW - Traveling maintainer problem
UR - http://www.scopus.com/inward/record.url?scp=85134299191&partnerID=8YFLogxK
U2 - 10.1016/j.ejor.2022.06.044
DO - 10.1016/j.ejor.2022.06.044
M3 - Article
SN - 0377-2217
VL - 305
SP - 1141
EP - 1152
JO - European Journal of Operational Research
JF - European Journal of Operational Research
IS - 3
ER -