z-logo
Premium
Flexible Servers in Understaffed Tandem Lines
Author(s) -
Kırkızlar Eser,
Andradóttir Sigrún,
Ayhan Hayriye
Publication year - 2012
Publication title -
production and operations management
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 3.279
H-Index - 110
eISSN - 1937-5956
pISSN - 1059-1478
DOI - 10.1111/j.1937-5956.2011.01295.x
Subject(s) - heuristics , server , throughput , computer science , flexibility (engineering) , mathematical optimization , distributed computing , service (business) , markov process , computer network , mathematics , operating system , statistics , economy , economics , wireless
We study the dynamic assignment of cross‐trained servers to stations in understaffed lines with finite buffers. Our objective is to maximize the production rate. We identify optimal server assignment policies for systems with three stations, two servers, different flexibility structures, and either deterministic service times and arbitrary buffers or exponential service times and small buffers. We use these policies to develop server assignment heuristics for Markovian systems with larger buffer sizes that appear to yield near‐optimal throughput. In the deterministic setting, we prove that the best possible production rate with full server flexibility and infinite buffers can be attained with partial flexibility and zero buffers, and we identify the critical skills required to achieve this goal. We then present numerical results showing that these critical skills, employed with an effective server assignment policy, also yield near‐optimal throughput in the Markovian setting, even for small buffer sizes. Thus, our results suggest that partial flexibility is sufficient for near‐optimal performance, and that flexibility structures that are effective for deterministic and infinite‐buffered systems are also likely to perform well for finite‐buffered stochastic systems.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here