Abstract
With increasing product complexities in mass customisation in the automotive industry, the downsides of conventional production concepts like flow production get more pronounced. Their inability to deal with cycle time losses adequately opens up possibilities for new concepts like matrix-structured production (MSP). Due to the immanent dynamics of matrix-structured production, control concept like takt binding or control stands are no longer sufficient to achieve near-optimal performance. The application of Reinforcement Learning (RL) to solve this problem emerged in the recent years. In particular, routing and dispatching tasks have been solved by applying RL. As both tasks influence each other's performance, a combined RL approach is developed. Therefore, a car body construction is simulated to test different modelled Markov processes, algorithms, and rewards. The new approach is validated against common heuristics regarding logistic performance and relevant metrics for operating autonomous guided vehicle fleets. For this, RL systems are designed and compared. The combined approach of production control in terms of dispatching jobs and routing autonomous guided vehicles achieved equivalent performance to heuristics. Still, it excelled in fleet operation metrics, like reduced live or deadlocks.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have