Markov decision processes in practice /

Saved in:
Bibliographic Details
Imprint:Cham, Switzerland : Springer, [2017]
©2017
Description:1 online resource
Language:English
Series:International series in operations research & management science ; volume 248
International series in operations research & management science ; v. 248.
Subject:
Format: E-Resource Book
URL for this record:http://pi.lib.uchicago.edu/1001/cat/bib/11272361
Hidden Bibliographic Details
Other authors / contributors:Boucherie, R. J. (Richard J.), 1964- editor.
Dijk, N. M. van, editor.
ISBN:9783319477664
3319477668
9783319477640
3319477641
Digital file characteristics:text file PDF
Notes:Includes bibliographical references.
Online resource, title from PDF title page (EBSCO, viewed March 18, 2017).
Summary:This book presents classical Markov Decision Processes (MDP) for real-life applications and optimization. MDP allows users to develop and formally support approximate and simple decision rules, and this book showcases state-of-the-art applications in which MDP was key to the solution approach. The book is divided into six parts. Part 1 is devoted to the state-of-the-art theoretical foundation of MDP, including approximate methods such as policy improvement, successive approximation and infinite state spaces as well as an instructive chapter on Approximate Dynamic Programming. It then continues with five parts of specific and non-exhaustive application areas. Part 2 covers MDP healthcare applications, which includes different screening procedures, appointment scheduling, ambulance scheduling and blood management. Part 3 explores MDP modeling within transportation. This ranges from public to private transportation, from airports and traffic lights to car parking or charging your electric car. Part 4 contains three chapters that illustrates the structure of approximate policies for production or manufacturing structures. In Part 5, communications is highlighted as an important application area for MDP. It includes Gittins indices, down-to-earth call centers and wireless sensor networks. Finally Part 6 is dedicated to financial modeling, offering an instructive review to account for financial portfolios and derivatives under proportional transactional costs. The MDP applications in this book illustrate a variety of both standard and non-standard aspects of MDP modeling and its practical use. This book should appeal to readers for practitioning, academic research and educational purposes, with a background in, among others, operations research, mathematics, computer science, and industrial engineering.
Other form:Print version: Markov decision processes in practice. Cham, Switzerland : Springer, [2017] 3319477641 9783319477640
Standard no.:10.1007/978-3-319-47766-4
10.1007/978-3-319-47

MARC

LEADER 00000cam a2200000Ii 4500
001 11272361
005 20210625184033.9
006 m o d
007 cr cnu|||unuuu
008 170314s2017 sz ob 000 0 eng d
015 |a GBB8J4728  |2 bnb 
016 7 |a 019091146  |2 Uk 
019 |a 975928413  |a 976133801  |a 976240432  |a 978414155  |a 978750711  |a 979006523  |a 979315861  |a 979439346  |a 984868719  |a 1012018949  |a 1066585885  |a 1105183446  |a 1112588243  |a 1112914002  |a 1113159755  |a 1122813052  |a 1127134975  |a 1134830156  |a 1156121381 
020 |a 9783319477664  |q (electronic bk.) 
020 |a 3319477668  |q (electronic bk.) 
020 |z 9783319477640 
020 |z 3319477641 
024 7 |a 10.1007/978-3-319-47766-4  |2 doi 
024 8 |a 10.1007/978-3-319-47 
035 |a (OCoLC)975486905  |z (OCoLC)975928413  |z (OCoLC)976133801  |z (OCoLC)976240432  |z (OCoLC)978414155  |z (OCoLC)978750711  |z (OCoLC)979006523  |z (OCoLC)979315861  |z (OCoLC)979439346  |z (OCoLC)984868719  |z (OCoLC)1012018949  |z (OCoLC)1066585885  |z (OCoLC)1105183446  |z (OCoLC)1112588243  |z (OCoLC)1112914002  |z (OCoLC)1113159755  |z (OCoLC)1122813052  |z (OCoLC)1127134975  |z (OCoLC)1134830156  |z (OCoLC)1156121381 
037 |a com.springer.onix.9783319477664  |b Springer Nature 
040 |a N$T  |b eng  |e rda  |e pn  |c N$T  |d IDEBK  |d N$T  |d OCLCO  |d EBLCP  |d YDX  |d NJR  |d UAB  |d OCLCF  |d IOG  |d AZU  |d UPM  |d MERER  |d ESU  |d VT2  |d Z5A  |d OCLCQ  |d JBG  |d IAD  |d ICW  |d ICN  |d OHI  |d OTZ  |d OCLCQ  |d IAS  |d OCLCQ  |d U3W  |d JG0  |d CAUOI  |d KSU  |d UKMGB  |d WYU  |d BRX  |d UKAHL  |d LQU  |d OCLCQ  |d ERF  |d UKBTH  |d LEATE  |d OCLCQ  |d SRU 
049 |a MAIN 
050 4 |a QA274.7 
072 7 |a MAT  |x 003000  |2 bisacsh 
072 7 |a MAT  |x 029000  |2 bisacsh 
072 7 |a KJT  |2 bicssc 
072 7 |a KJMD  |2 bicssc 
245 0 0 |a Markov decision processes in practice /  |c Richard J. Boucherie, Nico M. van Dijk. 
264 1 |a Cham, Switzerland :  |b Springer,  |c [2017] 
264 4 |c ©2017 
300 |a 1 online resource 
336 |a text  |b txt  |2 rdacontent 
337 |a computer  |b c  |2 rdamedia 
338 |a online resource  |b cr  |2 rdacarrier 
347 |a text file  |b PDF  |2 rda 
490 1 |a International series in operations research & management science ;  |v volume 248 
504 |a Includes bibliographical references. 
588 0 |a Online resource, title from PDF title page (EBSCO, viewed March 18, 2017). 
505 0 |a Foreword; Preface; Part I: General Theory; Part II: Healthcare; Part III: Transportation; Part IV: Production; Part V: Communications; Part VI: Financial Modeling; Summarizing; Acknowledgments; Contents; List of Contributors ; Part I General Theory; 1 One-Step Improvement Ideas and Computational Aspects; 1.1 Introduction; 1.2 The Average-Cost Markov Decision Model; 1.2.1 The Concept of Relative Values; 1.2.2 The Policy-Improvement Step; 1.2.3 The Odoni Bounds for Value Iteration; 1.3 Tailor-Made Policy-Iteration Algorithm; 1.3.1 A Queueing Control Problem with a Variable Service Rate. 
505 8 |a 1.4 One-Step Policy Improvement for Suboptimal Policies; 1.4.1 Dynamic Routing of Customers to Parallel Queues; 1.5 One-Stage-Look-Ahead Rule in Optimal Stopping; 1.5.1 Devil's Penny Problem; 1.5.2 A Game of Dropping Balls into Bins; 1.5.3 The Chow-Robbins Game; References; 2 Value Function Approximation in Complex Queueing Systems; 2.1 Introduction; 2.2 Difference Calculus for Markovian Birth-Death Systems; 2.3 Value Functions for Queueing Systems; 2.3.1 The M/Cox(r)/1 Queue; 2.3.2 Special Cases of the M/Cox(r)/1 Queue; 2.3.3 The M/M/s Queue; 2.3.4 The Blocking Costs in an M/M/s/s Queue. 
505 8 |a 2.3.5 Priority Queues; 2.4 Application: Routing to Parallel Queues; 2.5 Application: Dynamic Routing in Multiskill Call Centers; 2.6 Application: A Controlled Polling System; References; 3 Approximate Dynamic Programming by Practical Examples; 3.1 Introduction; 3.2 The Nomadic Trucker Example; 3.2.1 Problem Introduction; 3.2.2 MDP Model; 3.2.2.1 State; 3.2.2.2 Decision; 3.2.2.3 Costs; 3.2.2.4 New Information and Transition Function; 3.2.2.5 Solution; 3.2.3 Approximate Dynamic Programming; 3.2.3.1 Post-decision State; 3.2.3.2 Forward Dynamic Programming; 3.2.3.3 Value Function Approximation. 
505 8 |a 3.3 A Freight Consolidation Example; 3.3.1 Problem Introduction; 3.3.2 MDP Model; 3.3.2.1 State; 3.3.2.2 Decision; 3.3.2.3 Costs; 3.3.2.4 New Information and Transition Function; 3.3.2.5 Solution; 3.3.3 Approximate Dynamic Programming; 3.3.3.1 Post-decision State; 3.3.3.2 Forward Dynamic Programming; 3.3.3.3 Value Function Approximation; 3.4 A Healthcare Example; 3.4.1 Problem Introduction; 3.4.2 MDP Model; 3.4.2.1 State; 3.4.2.2 Decision; 3.4.2.3 Costs; 3.4.2.4 New Information and Transition Function; 3.4.2.5 Solution; 3.4.3 Approximate Dynamic Programming; 3.4.3.1 Post-decision State. 
505 8 |a 3.4.3.2 Forward Dynamic Programming; 3.4.3.3 Value Function Approximation; 3.5 What's More; 3.5.1 Policies; 3.5.2 Value Function Approximations; 3.5.3 Exploration vs Exploitation; Appendix; References; 4 Server Optimization of Infinite Queueing Systems; 4.1 Introduction; 4.2 Basic Definition and Notations; 4.3 Motivating Examples; 4.3.1 Optimization of a Queueing System with Two Different Servers; 4.3.2 Optimization of a Computational System with Power Saving Mode; 4.3.3 Structural Properties of These Motivating Examples; 4.4 Theoretical Background; 4.4.1 Subset Measures in Markov Chains. 
520 |a This book presents classical Markov Decision Processes (MDP) for real-life applications and optimization. MDP allows users to develop and formally support approximate and simple decision rules, and this book showcases state-of-the-art applications in which MDP was key to the solution approach. The book is divided into six parts. Part 1 is devoted to the state-of-the-art theoretical foundation of MDP, including approximate methods such as policy improvement, successive approximation and infinite state spaces as well as an instructive chapter on Approximate Dynamic Programming. It then continues with five parts of specific and non-exhaustive application areas. Part 2 covers MDP healthcare applications, which includes different screening procedures, appointment scheduling, ambulance scheduling and blood management. Part 3 explores MDP modeling within transportation. This ranges from public to private transportation, from airports and traffic lights to car parking or charging your electric car. Part 4 contains three chapters that illustrates the structure of approximate policies for production or manufacturing structures. In Part 5, communications is highlighted as an important application area for MDP. It includes Gittins indices, down-to-earth call centers and wireless sensor networks. Finally Part 6 is dedicated to financial modeling, offering an instructive review to account for financial portfolios and derivatives under proportional transactional costs. The MDP applications in this book illustrate a variety of both standard and non-standard aspects of MDP modeling and its practical use. This book should appeal to readers for practitioning, academic research and educational purposes, with a background in, among others, operations research, mathematics, computer science, and industrial engineering. 
650 0 |a Markov processes.  |0 http://id.loc.gov/authorities/subjects/sh85081369 
650 0 |a Decision making.  |0 http://id.loc.gov/authorities/subjects/sh85036199 
650 7 |a Operational research.  |2 bicssc 
650 7 |a Probability & statistics.  |2 bicssc 
650 7 |a MATHEMATICS  |x Applied.  |2 bisacsh 
650 7 |a MATHEMATICS  |x Probability & Statistics  |x General.  |2 bisacsh 
650 7 |a Decision making.  |2 fast  |0 (OCoLC)fst00889035 
650 7 |a Markov processes.  |2 fast  |0 (OCoLC)fst01010347 
655 0 |a Electronic books. 
655 4 |a Electronic books. 
700 1 |a Boucherie, R. J.  |q (Richard J.),  |d 1964-  |e editor.  |0 http://id.loc.gov/authorities/names/no93017962 
700 1 |a Dijk, N. M. van,  |e editor.  |0 http://id.loc.gov/authorities/names/n85204817 
776 0 8 |i Print version:  |t Markov decision processes in practice.  |d Cham, Switzerland : Springer, [2017]  |z 3319477641  |z 9783319477640  |w (OCoLC)959033672 
830 0 |a International series in operations research & management science ;  |v v. 248.  |0 http://id.loc.gov/authorities/names/n95066827 
903 |a HeVa 
929 |a oclccm 
999 f f |i cf3c5110-7114-58e0-bff8-bafc071e61fa  |s c57fc54a-743e-547b-a30d-91f2e85efdbe 
928 |t Library of Congress classification  |a QA274.7  |l Online  |c UC-FullText  |u https://link.springer.com/10.1007/978-3-319-47766-4  |z Springer Nature  |g ebooks  |i 12544699