3,155
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Reinforcement learning based optimal decision making towards product lifecycle sustainability

ORCID Icon, ORCID Icon & ORCID Icon
Pages 1269-1296 | Received 04 Sep 2020, Accepted 31 Dec 2021, Published online: 31 Jan 2022
 

ABSTRACT

Artificial intelligence (AI) has been widely used in robotics, automation, finance, healthcare, etc. However, using AI for decision-making in sustainable product lifecycle operations is still challenging. One major challenge relates to the scarcity and uncertainties of data across the product lifecycle. This paper aims to develop a method that can adopt the most suitable AI techniques to support decision-making for sustainable operations based on the available lifecycle data. It identifies the key lifecycle stages in which AI, especially reinforcement learning (RL), can support decision-making. Then, a generalised procedure of using RL for decision support is proposed based on available lifecycle data, such as operation and maintenance data. The method has been validated in a case study of an international vehicle manufacturer, combined with modelling and simulation. The case study demonstrates the effectiveness of the method and identifies that RL is the current most appropriate method for maintenance scheduling based on limited available lifecycle data. This paper contributes to knowledge by demonstrating an empirically grounded industrial case using RL to optimise decision-making for improved product lifecycle sustainability by effectively prolonging the product lifetime and reducing environmental impact.

Acknowledgments

A part of this paper (mainly the case study and the appendix) is based on excerpts from an unpublished internal report authored by P. Doherty, O. Andersson and J. Kvarnström (Linköping University), as part of the Vinnova project stated in the funding.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Correction Statement

This article has been republished with minor changes. These changes do not impact the academic content of the article.

Additional information

Funding

This work was supported by the Vinnova [2017-01649].