Academic literature on the topic 'Approximate dynamic programming'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Approximate dynamic programming.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Approximate dynamic programming"
Powell, Warren B. "Perspectives of approximate dynamic programming." Annals of Operations Research 241, no. 1-2 (February 7, 2012): 319–56. http://dx.doi.org/10.1007/s10479-012-1077-6.
Full textKulkarni, Sameer, Rajesh Ganesan, and Lance Sherry. "Dynamic Airspace Configuration Using Approximate Dynamic Programming." Transportation Research Record: Journal of the Transportation Research Board 2266, no. 1 (January 2012): 31–37. http://dx.doi.org/10.3141/2266-04.
Full textde Farias, D. P., and B. Van Roy. "The Linear Programming Approach to Approximate Dynamic Programming." Operations Research 51, no. 6 (December 2003): 850–65. http://dx.doi.org/10.1287/opre.51.6.850.24925.
Full textLogé, Frédéric, Erwan Le Pennec, and Habiboulaye Amadou-Boubacar. "Intelligent Questionnaires Using Approximate Dynamic Programming." i-com 19, no. 3 (December 1, 2020): 227–37. http://dx.doi.org/10.1515/icom-2020-0022.
Full textRyzhov, Ilya O., Martijn R. K. Mes, Warren B. Powell, and Gerald van den Berg. "Bayesian Exploration for Approximate Dynamic Programming." Operations Research 67, no. 1 (January 2019): 198–214. http://dx.doi.org/10.1287/opre.2018.1772.
Full textMaxwell, Matthew S., Mateo Restrepo, Shane G. Henderson, and Huseyin Topaloglu. "Approximate Dynamic Programming for Ambulance Redeployment." INFORMS Journal on Computing 22, no. 2 (May 2010): 266–81. http://dx.doi.org/10.1287/ijoc.1090.0345.
Full textCoşgun, Özlem, Ufuk Kula, and Cengiz Kahraman. "Markdown Optimization via Approximate Dynamic Programming." International Journal of Computational Intelligence Systems 6, no. 1 (February 2013): 64–78. http://dx.doi.org/10.1080/18756891.2013.754181.
Full textEl-Rayes, Khaled, and Hisham Said. "Dynamic Site Layout Planning Using Approximate Dynamic Programming." Journal of Computing in Civil Engineering 23, no. 2 (March 2009): 119–27. http://dx.doi.org/10.1061/(asce)0887-3801(2009)23:2(119).
Full textLee, Jay H., and Wee Chin Wong. "Approximate dynamic programming approach for process control." IFAC Proceedings Volumes 42, no. 11 (2009): 26–35. http://dx.doi.org/10.3182/20090712-4-tr-2008.00006.
Full textMcGrew, James S., Jonathon P. How, Brian Williams, and Nicholas Roy. "Air-Combat Strategy Using Approximate Dynamic Programming." Journal of Guidance, Control, and Dynamics 33, no. 5 (September 2010): 1641–54. http://dx.doi.org/10.2514/1.46815.
Full textDissertations / Theses on the topic "Approximate dynamic programming"
Sadiq, Mohammad. "Approximate Dynamic Programming Methods in HEVs." Thesis, KTH, Maskinkonstruktion (Inst.), 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-182762.
Full textElhybridfordon (HEV) har ökat i popularitet över hela världen pga sin låga bränsleförbrukning, vilket har minskat efterfrågan på olja. Detta gynnar i hög grad miljön, eftersom detta leder till mindre utsläpp och följaktligen en lägre växthuseffekt. Därför pågår aktiv forskning inom detta område med en ökad efterfrågan på nya och bättre strategier för bränsleförbrukning. Många olika metoder för energihantering av HEV använder en särskild metod; dynamisk programmering. Dynamisk programmering ger ett optimalt globalt resultat men på bekostnad av längre beräkningstider. Den mest använda metoden för att motverka denna typ av problematik i högdimensionella system är Approximate Dynamic Programming (ADP). Denna avhandling undersöker och beskriver litteraturen på de olika metoderna för ADP tillämpade på HEV samt en genomförandefas som visar en minskning av beräkningstiden för ett HEV-problem gällande energihantering.
Vyzas, Elias. "Approximate dynamic programming for some queueing problems." Thesis, Massachusetts Institute of Technology, 1997. http://hdl.handle.net/1721.1/10282.
Full textIncludes bibliographical references (p. 81-82).
by Elias Vyzas.
M.S.
Sauré, Antoine. "Approximate dynamic programming methods for advance patient scheduling." Thesis, University of British Columbia, 2012. http://hdl.handle.net/2429/43448.
Full textChild, Christopher H. T. "Approximate dynamic programming with parallel stochastic planning operators." Thesis, City University London, 2011. http://openaccess.city.ac.uk/1109/.
Full textLiu, Ning. "Approximate dynamic programming algorithms for production-planning problems." Thesis, Wichita State University, 2013. http://hdl.handle.net/10057/10636.
Full textThesis (M.S.)--Wichita State University, College of Engineering, Dept. of Industrial and Manufacturing Engineering
Demir, Ramazan. "An approximate dynamic programming approach to discrete optimization." Thesis, Massachusetts Institute of Technology, 2000. http://hdl.handle.net/1721.1/9137.
Full textIncludes bibliographical references (leaves 181-189).
We develop Approximate Dynamic Programming (ADP) methods to integer programming problems. We describe and investigate parametric, nonparametric and base-heuristic learning approaches to approximate the value function in order to break the curse of dimensionality. Through an extensive computational study we illustrate that our ADP approach to integer programming competes successfully with existing methodologies including state of art commercial packages like CPLEX. Our benchmarks for comparison are solution quality, running time and robustness (i.e., small deviations in the computational resources such as running time for varying instances of same size). In this thesis, we particularly focus on knapsack problems and the binary integer programming problem. We explore an integrated approach to solve discrete optimization problems by unifying optimization techniques with statistical learning. Overall, this research illustrates that the ADP is a promising technique by providing near-optimal solutions within reasonable amount of computation time especially for large scale problems with thousands of variables and constraints. Thus, Approximate Dynamic Programming can be considered as a new alternative to existing approximate methods for discrete optimization problems.
by Ramazan Demir.
Ph.D.
Cai, C. "Adaptive traffic signal control using approximate dynamic programming." Thesis, University College London (University of London), 2010. http://discovery.ucl.ac.uk/20164/.
Full textNadarajah, Selvaprabu. "Approximate Dynamic Programming for Commodity and Energy Merchant Operations." Research Showcase @ CMU, 2014. http://repository.cmu.edu/dissertations/350.
Full textBethke, Brett (Brett M. ). "Kernel-based approximate dynamic programming using Bellman residual elimination." Thesis, Massachusetts Institute of Technology, 2010. http://hdl.handle.net/1721.1/57544.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student submitted PDF version of thesis.
Includes bibliographical references (p. 207-221).
Many sequential decision-making problems related to multi-agent robotic systems can be naturally posed as Markov Decision Processes (MDPs). An important advantage of the MDP framework is the ability to utilize stochastic system models, thereby allowing the system to make sound decisions even if there is randomness in the system evolution over time. Unfortunately, the curse of dimensionality prevents most MDPs of practical size from being solved exactly. One main focus of the thesis is on the development of a new family of algorithms for computing approximate solutions to large-scale MDPs. Our algorithms are similar in spirit to Bellman residual methods, which attempt to minimize the error incurred in solving Bellman's equation at a set of sample states. However, by exploiting kernel-based regression techniques (such as support vector regression and Gaussian process regression) with nondegenerate kernel functions as the underlying cost-to-go function approximation architecture, our algorithms are able to construct cost-to-go solutions for which the Bellman residuals are explicitly forced to zero at the sample states. For this reason, we have named our approach Bellman residual elimination (BRE). In addition to developing the basic ideas behind BRE, we present multi-stage and model-free extensions to the approach. The multistage extension allows for automatic selection of an appropriate kernel for the MDP at hand, while the model-free extension can use simulated or real state trajectory data to learn an approximate policy when a system model is unavailable.
(cont.) We present theoretical analysis of all BRE algorithms proving convergence to the optimal policy in the limit of sampling the entire state space, and show computational results on several benchmark problems. Another challenge in implementing control policies based on MDPs is that there may be parameters of the system model that are poorly known and/or vary with time as the system operates. System performance can suer if the model used to compute the policy differs from the true model. To address this challenge, we develop an adaptive architecture that allows for online MDP model learning and simultaneous re-computation of the policy. As a result, the adaptive architecture allows the system to continuously re-tune its control policy to account for better model information 3 obtained through observations of the actual system in operation, and react to changes in the model as they occur. Planning in complex, large-scale multi-agent robotic systems is another focus of the thesis. In particular, we investigate the persistent surveillance problem, in which one or more unmanned aerial vehicles (UAVs) and/or unmanned ground vehicles (UGVs) must provide sensor coverage over a designated location on a continuous basis. This continuous coverage must be maintained even in the event that agents suer failures over the course of the mission. The persistent surveillance problem is pertinent to a number of applications, including search and rescue, natural disaster relief operations, urban traffic monitoring, etc.
(cont.) Using both simulations and actual flight experiments conducted in the MIT RAVEN indoor flight facility, we demonstrate the successful application of the BRE algorithms and the adaptive MDP architecture in achieving high mission performance despite the random occurrence of failures. Furthermore, we demonstrate performance benefits of our approach over a deterministic planning approach that does not account for these failures.
by Brett M. Bethke.
Ph.D.
Valenti, Mario J. (Mario James) 1976. "Approximate dynamic programming with applications in multi-agent systems." Thesis, Massachusetts Institute of Technology, 2007. http://hdl.handle.net/1721.1/40330.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
MIT Institute Archives copy: contains CDROM of thesis in .pdf format.
Includes bibliographical references (p. 151-161).
This thesis presents the development and implementation of approximate dynamic programming methods used to manage multi-agent systems. The purpose of this thesis is to develop an architectural framework and theoretical methods that enable an autonomous mission system to manage real-time multi-agent operations. To meet this goal, we begin by discussing aspects of the real-time multi-agent mission problem. Next, we formulate this problem as a Markov Decision Process (MDP) and present a system architecture designed to improve mission-level functional reliability through system self-awareness and adaptive mission planning. Since most multi-agent mission problems are computationally difficult to solve in real-time, approximation techniques are needed to find policies for these large-scale problems. Thus, we have developed theoretical methods used to find feasible solutions to large-scale optimization problems. More specifically, we investigate methods designed to automatically generate an approximation to the cost-to-go function using basis functions for a given MDP. Next, these these techniques are used by an autonomous mission system to manage multi-agent mission scenarios. Simulation results using these methods are provided for a large-scale mission problem. In addition, this thesis presents the implementation of techniques used to manage autonomous unmanned aerial vehicles (UAVs) performing persistent surveillance operations. We present an indoor multi-vehicle testbed called RAVEN (Real-time indoor Autonomous Vehicle test ENvironment) that was developed to study long-duration missions in a controlled environment.
(cont.) The RAVEN's design allows researchers to focus on high-level tasks by autonomously managing the platform's realistic air and ground vehicles during multi-vehicle operations, thus promoting the rapid prototyping of UAV technologies by flight testing new vehicle configurations and algorithms without redesigning vehicle hardware. Finally, using the RAVEN, we present flight test results from autonomous, extended mission tests using the technologies developed in this thesis. Flight results from a 24 hr, fully-autonomous air vehicle flight-recharge test and an autonomous, multi-vehicle extended mission test using small, electric-powered air vehicles are provided.
by Mario J. Valenti.
Ph.D.
Books on the topic "Approximate dynamic programming"
Powell, Warren B. Approximate Dynamic Programming. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2011. http://dx.doi.org/10.1002/9781118029176.
Full textUlmer, Marlin Wolf. Approximate Dynamic Programming for Dynamic Vehicle Routing. Cham: Springer International Publishing, 2017. http://dx.doi.org/10.1007/978-3-319-55511-9.
Full textPowell, Warren B. Approximate dynamic programming: Solving the curses of dimensionality. 2nd ed. Hoboken, N.J: J. Wiley & Sons, 2011.
Find full textApproximate dynamic programming: Solving the curses of dimensionality. Hoboken, NJ: J. Wiley & Sons, 2007.
Find full textPowell, Warren B. Approximate dynamic programming: Solving the curses of dimensionality. Hoboken, NJ: Wiley-Interscience, 2007.
Find full textLewis, Frank L., and Derong Liu, eds. Reinforcement Learning and Approximate Dynamic Programming for Feedback Control. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2012. http://dx.doi.org/10.1002/9781118453988.
Full textThomas, L. C. Approximate solutions of moving target search problems using dynamic programming. Edinburgh: University of Edinburgh. Department of Business Studies, 1987.
Find full textIEEE, International Symposium on Approximate Dynamic Programming and Reinforcement Learning (1st 2007 Honolulu Hawaii). 2007 IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning: Honolulu, HI, 1-5 April 2007. Piscataway, NJ: IEEE, 2007.
Find full textIEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning (1st 2007 Honolulu, Hawaii). 2007 IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning: Honolulu, HI, 1-5 April 2007. Piscataway, NJ: IEEE, 2007.
Find full textIEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning (1st 2007 Honolulu, Hawaii). 2007 IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning: Honolulu, HI, 1-5 April 2007. Piscataway, NJ: IEEE, 2007.
Find full textBook chapters on the topic "Approximate dynamic programming"
Fu, Michael C. "Approximate Dynamic Programming." In Encyclopedia of Operations Research and Management Science, 73–77. Boston, MA: Springer US, 2013. http://dx.doi.org/10.1007/978-1-4419-1153-7_1189.
Full textKakas, Antonis C., David Cohn, Sanjoy Dasgupta, Andrew G. Barto, Gail A. Carpenter, Stephen Grossberg, Geoffrey I. Webb, et al. "Approximate Dynamic Programming." In Encyclopedia of Machine Learning, 39. Boston, MA: Springer US, 2011. http://dx.doi.org/10.1007/978-0-387-30164-8_26.
Full textKamalapurkar, Rushikesh, Patrick Walters, Joel Rosenfeld, and Warren Dixon. "Approximate Dynamic Programming." In Communications and Control Engineering, 17–42. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-78384-0_2.
Full textMunos, Rémi. "Approximate Dynamic Programming." In Markov Decision Processes in Artificial Intelligence, 67–98. Hoboken, NJ USA: John Wiley & Sons, Inc., 2013. http://dx.doi.org/10.1002/9781118557426.ch3.
Full textWerbos, Paul J. "Approximate Dynamic Programming (ADP)." In Encyclopedia of Systems and Control, 1–7. London: Springer London, 2020. http://dx.doi.org/10.1007/978-1-4471-5102-9_100096-1.
Full textWerbos, Paul J. "Approximate Dynamic Programming (ADP)." In Encyclopedia of Systems and Control, 76–82. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-44184-5_100096.
Full textJiang, Yu, and Zhong-Ping Jiang. "Robust Adaptive Dynamic Programming." In Reinforcement Learning and Approximate Dynamic Programming for Feedback Control, 281–302. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2013. http://dx.doi.org/10.1002/9781118453988.ch13.
Full textSeiffertt, John, and Donald C. Wunsch. "Approximate Dynamic Programming on Time Scales." In Evolutionary Learning and Optimization, 61–76. Berlin, Heidelberg: Springer Berlin Heidelberg, 2010. http://dx.doi.org/10.1007/978-3-642-03180-9_5.
Full textPowell, Warren B., and Ilya O. Ryzhov. "Optimal Learning and Approximate Dynamic Programming." In Reinforcement Learning and Approximate Dynamic Programming for Feedback Control, 410–31. Hoboken, NJ, USA: John Wiley & Sons, Inc., 2013. http://dx.doi.org/10.1002/9781118453988.ch18.
Full textVeksler, Olga. "Dynamic Programming for Approximate Expansion Algorithm." In Computer Vision – ECCV 2012, 850–63. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-33712-3_61.
Full textConference papers on the topic "Approximate dynamic programming"
Dyer, Martin. "Approximate counting by dynamic programming." In the thirty-fifth ACM symposium. New York, New York, USA: ACM Press, 2003. http://dx.doi.org/10.1145/780542.780643.
Full textO'Donoghue, Brendan, Yang Wang, and Stephen Boyd. "Min-max approximate dynamic programming." In Control (MSC). IEEE, 2011. http://dx.doi.org/10.1109/cacsd.2011.6044538.
Full textSummers, Tyler H., Konstantin Kunz, Nikolaos Kariotoglou, Maryam Kamgarpour, Sean Summers, and John Lygeros. "Approximate dynamic programming via sum of squares programming." In 2013 European Control Conference (ECC). IEEE, 2013. http://dx.doi.org/10.23919/ecc.2013.6669374.
Full textPreux, Philippe, Sertan Girgin, and Manuel Loth. "Feature discovery in approximate dynamic programming." In 2009 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL). IEEE, 2009. http://dx.doi.org/10.1109/adprl.2009.4927533.
Full textWang, Lin, Hui Peng, Hua-yong Zhu, and Lin-cheng Shen. "A Survey of Approximate Dynamic Programming." In 2009 International Conference on Intelligent Human-Machine Systems and Cybernetics. IEEE, 2009. http://dx.doi.org/10.1109/ihmsc.2009.222.
Full textDeisenroth, Marc P., Jan Peters, and Carl E. Rasmussen. "Approximate dynamic programming with Gaussian processes." In 2008 American Control Conference (ACC '08). IEEE, 2008. http://dx.doi.org/10.1109/acc.2008.4587201.
Full textKariotoglou, Nikolaos, Sean Summers, Tyler Summers, Maryam Kamgarpour, and John Lygeros. "Approximate dynamic programming for stochastic reachability." In 2013 European Control Conference (ECC). IEEE, 2013. http://dx.doi.org/10.23919/ecc.2013.6669603.
Full textZhao, Dongbin, Jianqiang Yi, and Derong Liu. "Particle Swarn Optimized Adaptive Dynamic Programming." In 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning. IEEE, 2007. http://dx.doi.org/10.1109/adprl.2007.368166.
Full textAtkeson, Christopher G. "Randomly Sampling Actions In Dynamic Programming." In 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning. IEEE, 2007. http://dx.doi.org/10.1109/adprl.2007.368187.
Full textSutter, Tobias, Angeliki Kamoutsi, Peyman Mohajerin Esfahani, and John Lygeros. "Data-driven approximate dynamic programming: A linear programming approach." In 2017 IEEE 56th Annual Conference on Decision and Control (CDC). IEEE, 2017. http://dx.doi.org/10.1109/cdc.2017.8264426.
Full textReports on the topic "Approximate dynamic programming"
Nachmani, Gil. Minimum-Energy Flight Paths for UAVs Using Mesoscale Wind Forecasts and Approximate Dynamic Programming. Fort Belvoir, VA: Defense Technical Information Center, December 2007. http://dx.doi.org/10.21236/ada475720.
Full text