Abstract

In this chapter, the system defined by a subcontractor, which processes orders generated from multiple customers, is cast into the dynamic single-machine sequencing framework. The subcontractor is modeled as a production/inventory facility that consists of a single, failure-free server and an input buffer with infinite capacity. Jobs arrive continuously to the system at random intervals and have stochastic processing times and due dates. A setup is required when switching from one job type to another. The behaviour of the system, when operating under various rescheduling policies and job dispatching schemes, is investigated using discrete-event simulation in a series of experiments. A reinforcement learning-based approach is proposed for dynamic dispatching rule selection under the completely reactive rescheduling policy. The reinforcement learning controller is applied to two simulation cases with the aim of generating state-dispatching rule mappings that minimize mean tardiness and mean lateness.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.