default search action
ADPRL 2014: Orlando, FL, USA
- 2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL 2014, Orlando, FL, USA, December 9-12, 2014. IEEE 2014, ISBN 978-1-4799-4553-5
- Wei Sun, Evangelos A. Theodorou, Panagiotis Tsiotras:
Continuous-time differential dynamic programming with terminal constraints. 1-6 - Oktay Arslan, Evangelos A. Theodorou, Panagiotis Tsiotras:
Information-theoretic stochastic optimal control via incremental sampling-based algorithms. 1-8 - Timothé Collet, Olivier Pietquin:
Active learning for classification: An optimistic approach. 1-8 - Xiangnan Zhong, Zhen Ni, Yufei Tang, Haibo He:
Data-driven partially observable dynamic processes using adaptive dynamic programming. 1-8 - Eugene A. Feinberg, Pavlo O. Kasyanov, Michael Z. Zgurovsky:
Convergence of value iterations for total-cost MDPs and POMDPs with general state and action sets. 1-8 - Lei Liu, Zhanshan Wang, Zhengwei Shen:
Neural-network-based adaptive dynamic surface control for MIMO systems with unknown hysteresis. 1-6 - Balázs Csanád Csáji, András Kovács, József Váncza:
Adaptive aggregated predictions for renewable energy systems. 1-8 - Ali Heydari:
Theoretical analysis of a reinforcement learning based switching scheme. 1-6 - Xiaohong Cui, Yanhong Luo, Huaguang Zhang:
An adaptive dynamic programming algorithm to solve optimal control of uncertain nonlinear systems. 1-6 - Simon Haykin, Ashkan Amiri, Mehdi Fatemi:
Cognitive control in cognitive dynamic systems: A new way of thinking inspired by the brain. 1-7 - Daniel L. Elliott, Charles Anderson:
Using supervised training signals of observable state dynamics to speed-up and improve reinforcement learning. 1-8 - Yang Liu, Yanhong Luo, Huaguang Zhang:
Adaptive dynamic programming for discrete-time LQR optimal tracking control problems with unknown dynamics. 1-6 - Taishi Fujita, Toshimitsu Ushio:
Reinforcement learning-based optimal control considering L computation time delay of linear discrete-time systems. 1-6 - Hadrien Glaude, Olivier Pietquin, Cyrille Enderli:
Subspace identification for predictive state representation by nuclear norm minimization. 1-8 - Deon Garrett, Jordi Bieger, Kristinn R. Thórisson:
Tunable and generic problem instance generation for multi-objective reinforcement learning. 1-8 - Martin W. Allen, David Hahn, Douglas C. MacFarland:
Heuristics for multiagent reinforcement learning in decentralized decision problems. 1-8 - Madalina M. Drugan, Ann Nowé, Bernard Manderick:
Pareto Upper Confidence Bounds algorithms: An empirical study. 1-8 - Minwoo Lee, Charles W. Anderson:
Convergent reinforcement learning control with neural networks and continuous action search. 1-8 - Yuhai Hu, Boris Defourny:
Near-optimality bounds for greedy periodic policies with application to grid-level storage. 1-8 - Marco A. Wiering, Maikel Withagen, Madalina M. Drugan:
Model-based multi-objective reinforcement learning. 1-6 - Hao Xu, Sarangapani Jagannathan:
Model-free Q-learning over finite horizon for uncertain linear continuous-time systems. 1-6 - Avimanyu Sahoo, Hao Xu, Sarangapani Jagannathan:
Event-based optimal regulator design for nonlinear networked control systems. 1-8 - Li-Bing Wu, Dan Ye, Xin-Gang Zhao:
Adaptive fault identification for a class of nonlinear dynamic systems. 1-6 - Hengshuai Yao, Csaba Szepesvári, Bernardo Ávila Pires, Xinhua Zhang:
Pseudo-MDPs and factored linear action models. 1-9 - Qinglai Wei, Derong Liu, Guang Shi, Yu Liu, Qiang Guan:
Optimal self-learning battery control in smart residential grids by iterative Q-learning algorithm. 1-7 - Simone Parisi, Matteo Pirotta, Nicola Smacchia, Luca Bascetta, Marcello Restelli:
Policy gradient approaches for multi-objective sequential decision making: A comparison. 1-8 - Sumit Kumar Jha, Shubhendu Bhasin:
On-policy Q-learning for adaptive optimal control. 1-6 - Seyed Reza Ahmadzadeh, Petar Kormushev, Darwin G. Caldwell:
Multi-objective reinforcement learning for AUV thruster failure recovery. 1-8 - Vincent François-Lavet, Raphaël Fonteneau, Damien Ernst:
Using approximate dynamic programming for estimating the revenues of a hydrogen-based high-capacity storage device. 1-8 - Xiaofeng Lin, Qiang Ding, Weikai Kong, Chunning Song, Qingbao Huang:
Adaptive dynamic programming-based optimal tracking control for nonlinear systems using general value iteration. 1-6 - Haci Mehmet Guzey, Hao Xu, Sarangapani Jagannathan:
Neural network-based adaptive optimal consensus control of leaderless networked mobile robots. 1-6 - Lucian Busoniu, Rémi Munos, Elod Páll:
An analysis of optimistic, best-first search for minimax sequential decision making. 1-8 - Dominik Meyer, Rémy Degenne, Ahmed Omrane, Hao Shen:
Accelerated gradient temporal difference learning algorithms. 1-8 - Daniel R. Jiang, Thuy V. Pham, Warren B. Powell, Daniel F. Salas, Warren R. Scott:
A comparison of approximate dynamic programming techniques on benchmark energy storage problems: Does anything work? 1-8 - Yanhong Luo, Geyang Xiao:
ADP-based optimal control for a class of nonlinear discrete-time systems with inequality constraints. 1-5 - Regina Padmanabhan, Nader Meskin, Wassim M. Haddad:
Closed-loop control of anesthesia and mean arterial pressure using reinforcement learning. 1-8 - Abhijit Gosavi, Sajal K. Das, Susan L. Murray:
Beyond exponential utility functions: A variance-adjusted approach for risk-averse reinforcement learning. 1-8 - Yunpeng Pan, Evangelos A. Theodorou:
Nonparametric infinite horizon Kullback-Leibler stochastic control. 1-8 - Saba Q. Yahyaa, Madalina M. Drugan, Bernard Manderick:
Annealing-pareto multi-objective multi-armed bandit algorithm. 1-8 - Ahmad A. Al-Talabi, Howard M. Schwartz:
A two stage learning technique for dual learning in the pursuit-evasion differential game. 1-8 - Yuanheng Zhu, Dongbin Zhao:
A data-based online reinforcement learning algorithm with high-efficient exploration. 1-6 - Joschka Boedecker, Jost Tobias Springenberg, Jan Wülfing, Martin A. Riedmiller:
Approximate real-time optimal control based on sparse Gaussian process models. 1-8
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.