Markov Decision Processes (MDP) have been widely used as a framework for planning under uncertainty. They allow to compute optimal sequences of actions in order to achieve a given...
Decentralized partially observable Markov decision processes (Dec-POMDPs) constitute a generic and expressive framework for multiagent planning under uncertainty. However, plannin...
Frans A. Oliehoek, Shimon Whiteson, Matthijs T. J....
One of the most important policies adopted in inventory control is the (R,S) policy (also known as the “replenishment cycle” policy). Under the non-stationary demand assumption...
— In recent years, the research community introduced various methods for processing skyline queries in multidimensional databases. The skyline operator retrieves all objects bein...
We propose a novel algorithmic framework to solve an integrated planning and scheduling problem in supply chain management. This problem involves the integration of an inventory m...