researchr
explore
Tags
Journals
Conferences
Authors
Profiles
Groups
calendar
New Conferences
Events
Deadlines
search
search
You are not signed in
Sign in
Sign up
Links
Filter by Year
OR
AND
NOT
1
2009
2011
2013
2014
Filter by Tag
Filter by Author
[+]
OR
AND
NOT
1
Bart De Schutter
Boris Defourny
Damien Ernst
Derong Liu
Dongbin Zhao
Emanuel Todorov
Evangelos A. Theodorou
Frank L. Lewis
Hao Xu
Huaguang Zhang
Lucian Busoniu
Madalina M. Drugan
Marco A. Wiering
Olivier Pietquin
Robert Babuska
Rémi Munos
Sarangapani Jagannathan
Warren B. Powell
Xin Xu
Yanhong Luo
Filter by Top terms
[+]
OR
AND
NOT
1
action
adaptive
algorithm
algorithms
approximate
continuous
control
discrete
dynamic
horizon
learning
multi
nonlinear
optimal
policy
programming
reinforcement
systems
time
using
ADPRL (adprl)
Editions
Publications
Viewing Publication 1 - 100 from 153
2014
2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL 2014, Orlando, FL, USA, December 9-12, 2014
IEEE,
2014.
[doi]
Multi-objective reinforcement learning for AUV thruster failure recovery
Seyed Reza Ahmadzadeh
,
Petar Kormushev
,
Darwin G. Caldwell
.
adprl 2014
:
1-8
[doi]
A two stage learning technique for dual learning in the pursuit-evasion differential game
Ahmad A. Al-Talabi
,
Howard M. Schwartz
.
adprl 2014
:
1-8
[doi]
Heuristics for multiagent reinforcement learning in decentralized decision problems
Martin W. Allen
,
David Hahn
,
Douglas C. MacFarland
.
adprl 2014
:
1-8
[doi]
Information-theoretic stochastic optimal control via incremental sampling-based algorithms
Oktay Arslan
,
Evangelos A. Theodorou
,
Panagiotis Tsiotras
.
adprl 2014
:
1-8
[doi]
Approximate real-time optimal control based on sparse Gaussian process models
Joschka Boedecker
,
Jost Tobias Springenberg
,
Jan Wülfing
,
Martin A. Riedmiller
.
adprl 2014
:
1-8
[doi]
An analysis of optimistic, best-first search for minimax sequential decision making
Lucian Busoniu
,
Rémi Munos
,
Elod Pall
.
adprl 2014
:
1-8
[doi]
Active learning for classification: An optimistic approach
Timothe Collet
,
Olivier Pietquin
.
adprl 2014
:
1-8
[doi]
Adaptive aggregated predictions for renewable energy systems
Balázs Csanád Csáji
,
András Kovács
,
József Váncza
.
adprl 2014
:
1-8
[doi]
An adaptive dynamic programming algorithm to solve optimal control of uncertain nonlinear systems
Xiaohong Cui
,
Yanhong Luo
,
Huaguang Zhang
.
adprl 2014
:
1-6
[doi]
Pareto Upper Confidence Bounds algorithms: An empirical study
Madalina M. Drugan
,
Ann Nowé
,
Bernard Manderick
.
adprl 2014
:
1-8
[doi]
Using supervised training signals of observable state dynamics to speed-up and improve reinforcement learning
Daniel L. Elliott
,
Charles Anderson
.
adprl 2014
:
1-8
[doi]
Convergence of value iterations for total-cost MDPs and POMDPs with general state and action sets
Eugene A. Feinberg
,
Pavlo O. Kasyanov
,
Michael Z. Zgurovsky
.
adprl 2014
:
1-8
[doi]
Using approximate dynamic programming for estimating the revenues of a hydrogen-based high-capacity storage device
Vincent François-Lavet
,
Raphaël Fonteneau
,
Damien Ernst
.
adprl 2014
:
1-8
[doi]
Reinforcement learning-based optimal control considering L computation time delay of linear discrete-time systems
Taishi Fujita
,
Toshimitu Ushio
.
adprl 2014
:
1-6
[doi]
Tunable and generic problem instance generation for multi-objective reinforcement learning
Deon Garrett
,
Jordi Bieger
,
Kristinn R. Thórisson
.
adprl 2014
:
1-8
[doi]
Subspace identification for predictive state representation by nuclear norm minimization
Hadrien Glaude
,
Olivier Pietquin
,
Cyrille Enderli
.
adprl 2014
:
1-8
[doi]
Beyond exponential utility functions: A variance-adjusted approach for risk-averse reinforcement learning
Abhijit Gosavi
,
Sajal K. Das
,
Susan L. Murray
.
adprl 2014
:
1-8
[doi]
Neural network-based adaptive optimal consensus control of leaderless networked mobile robots
Haci Mehmet Guzey
,
Hao Xu
,
Sarangapani Jagannathan
.
adprl 2014
:
1-6
[doi]
Cognitive control in cognitive dynamic systems: A new way of thinking inspired by the brain
Simon Haykin
,
Ashkan Amiri
,
Mehdi Fatemi
.
adprl 2014
:
1-7
[doi]
Theoretical analysis of a reinforcement learning based switching scheme
Ali Heydari
.
adprl 2014
:
1-6
[doi]
Near-optimality bounds for greedy periodic policies with application to grid-level storage
Yuhai Hu
,
Boris Defourny
.
adprl 2014
:
1-8
[doi]
On-policy Q-learning for adaptive optimal control
Sumit Kumar Jha
,
Shubhendu Bhasin
.
adprl 2014
:
1-6
[doi]
A comparison of approximate dynamic programming techniques on benchmark energy storage problems: Does anything work?
Daniel R. Jiang
,
Thuy V. Pham
,
Warren B. Powell
,
Daniel F. Salas
,
Warren R. Scott
.
adprl 2014
:
1-8
[doi]
Convergent reinforcement learning control with neural networks and continuous action search
Minwoo Lee
,
Charles W. Anderson
.
adprl 2014
:
1-8
[doi]
Adaptive dynamic programming-based optimal tracking control for nonlinear systems using general value iteration
Xiaofeng Lin
,
Qiang Ding
,
Weikai Kong
,
Chunning Song
,
Qingbao Huang
.
adprl 2014
:
1-6
[doi]
Neural-network-based adaptive dynamic surface control for MIMO systems with unknown hysteresis
Lei Liu
,
Zhanshan Wang
,
Zhengwei Shen
.
adprl 2014
:
1-6
[doi]
Adaptive dynamic programming for discrete-time LQR optimal tracking control problems with unknown dynamics
Yang Liu
,
Yanhong Luo
,
Huaguang Zhang
.
adprl 2014
:
1-6
[doi]
ADP-based optimal control for a class of nonlinear discrete-time systems with inequality constraints
Yanhong Luo
,
Geyang Xiao
.
adprl 2014
:
1-5
[doi]
Accelerated gradient temporal difference learning algorithms
Dominik Meyer
,
Remy Degenne
,
Ahmed Omrane
,
Hao Shen
.
adprl 2014
:
1-8
[doi]
Closed-loop control of anesthesia and mean arterial pressure using reinforcement learning
Regina Padmanabhan
,
Nader Meskin
,
Wassim M. Haddad
.
adprl 2014
:
1-8
[doi]
Nonparametric infinite horizon Kullback-Leibler stochastic control
Yunpeng Pan
,
Evangelos A. Theodorou
.
adprl 2014
:
1-8
[doi]
Policy gradient approaches for multi-objective sequential decision making: A comparison
Simone Parisi
,
Matteo Pirotta
,
Nicola Smacchia
,
Luca Bascetta
,
Marcello Restelli
.
adprl 2014
:
1-8
[doi]
Event-based optimal regulator design for nonlinear networked control systems
Avimanyu Sahoo
,
Hao Xu
,
Sarangapani Jagannathan
.
adprl 2014
:
1-8
[doi]
Continuous-time differential dynamic programming with terminal constraints
Wei Sun
,
Evangelos A. Theodorou
,
Panagiotis Tsiotras
.
adprl 2014
:
1-6
[doi]
Optimal self-learning battery control in smart residential grids by iterative Q-learning algorithm
Qinglai Wei
,
Derong Liu
,
Guang Shi
,
Yu Liu
,
Qiang Guan
.
adprl 2014
:
1-7
[doi]
Model-based multi-objective reinforcement learning
Marco A. Wiering
,
Maikel Withagen
,
Madalina M. Drugan
.
adprl 2014
:
1-6
[doi]
Adaptive fault identification for a class of nonlinear dynamic systems
Li-bing Wu
,
Dan Ye
,
Xin-Gang Zhao
.
adprl 2014
:
1-6
[doi]
Model-free Q-learning over finite horizon for uncertain linear continuous-time systems
Hao Xu
,
Sarangapani Jagannathan
.
adprl 2014
:
1-6
[doi]
Annealing-pareto multi-objective multi-armed bandit algorithm
Saba Q. Yahyaa
,
Madalina M. Drugan
,
Bernard Manderick
.
adprl 2014
:
1-8
[doi]
Pseudo-MDPs and factored linear action models
Hengshuai Yao
,
Csaba Szepesvári
,
Bernardo Avila Pires
,
Xinhua Zhang
.
adprl 2014
:
1-9
[doi]
Data-driven partially observable dynamic processes using adaptive dynamic programming
Xiangnan Zhong
,
Zhen Ni
,
Yufei Tang
,
Haibo He
.
adprl 2014
:
1-8
[doi]
A data-based online reinforcement learning algorithm with high-efficient exploration
Yuanheng Zhu
,
Dongbin Zhao
.
adprl 2014
:
1-6
[doi]
2013
Proceedings of the 2013 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL 2013, IEEE Symposium Series on Computational Intelligence (SSCI), 16-19 April 2013, Singapore
IEEE,
2013.
[doi]
Exponential moving average Q-learning algorithm
M. D. Awheda
,
H. M. Schwartz
.
adprl 2013
:
31-38
[doi]
Reinforcement learning to train Ms. Pac-Man using higher-order action-relative inputs
Luuk Bom
,
Ruud Henken
,
Marco Wiering
.
adprl 2013
:
156-163
[doi]
Optimistic planning for continuous-action deterministic systems
Lucian Busoniu
,
A. Daniels
,
Rémi Munos
,
Robert Babuska
.
adprl 2013
:
69-76
[doi]
A combined hierarchical reinforcement learning based approach for multi-robot cooperative target searching in complex unknown environments
Yifan Cai
,
Simon X. Yang
,
Xin Xu
.
adprl 2013
:
52-59
[doi]
Optimistic planning for belief-augmented Markov Decision Processes
Raphaël Fonteneau
,
Lucian Busoniu
,
Rémi Munos
.
adprl 2013
:
77-84
[doi]
The second order temporal difference error for Sarsa(λ)
Qi-ming Fu
,
Quan Liu
,
Fei Xiao
,
Guixin Chen
.
adprl 2013
:
60-68
[doi]
On the coordination system for the dimensionality-reduced inputs of mario
Hisashi Handa
.
adprl 2013
:
170-176
[doi]
Local stability analysis of high-order recurrent neural networks with multi-step piecewise linear activation functions
Yujiao Huang
,
Huaguang Zhang
,
Dongsheng Yang
.
adprl 2013
:
1-5
[doi]
Optimized look-ahead trees: Extensions to large and continuous action spaces
T. Jung
,
D. Ernst
,
F. Maes
.
adprl 2013
:
85-92
[doi]
A reinforcement learning algorithm developed to model GenCo strategic bidding behavior in multidimensional and continuous state and action spaces
A. Y. F. Lau
,
Dipti Srinivasan
,
Thomas Reindl
.
adprl 2013
:
116-123
[doi]
Bias-corrected Q-learning to control max-operator bias in Q-learning
Donghun Lee
,
Boris Defourny
,
Warren B. Powell
.
adprl 2013
:
93-99
[doi]
Optimal control for a class of nonlinear systems with state delay based on Adaptive Dynamic Programming with ε-error bound
Xiaofeng Lin
,
Nuyun Cao
,
Yuzhang Lin
.
adprl 2013
:
177-182
[doi]
Exploring the relationship of reward and punishment in reinforcement learning
Robert Lowe
,
Tom Ziemke
.
adprl 2013
:
140-147
[doi]
An integrated design for intensified direct heuristic dynamic programming
Xiong Luo
,
Jennie Si
,
Yuchao Zhou
.
adprl 2013
:
183-190
[doi]
Scalarized multi-objective reinforcement learning: Novel design techniques
Kristof Van Moffaert
,
Madalina M. Drugan
,
Ann Nowé
.
adprl 2013
:
191-199
[doi]
Real-time tracking on adaptive critic design with uniformly ultimately bounded condition
Zhen Ni
,
Xiao Fang
,
Haibo He
,
Dongbin Zhao
,
Xin Xu
.
adprl 2013
:
39-46
[doi]
Adaptive optimal control for nonlinear discrete-time systems
Chunbin Qin
,
Huaguang Zhang
,
Yanhong Luo
.
adprl 2013
:
13-18
[doi]
A study on the efficiency of learning a robot controller in various environments
Sachiko Soga
,
Ichiro Kobayashi
.
adprl 2013
:
164-169
[doi]
Optimal control for a class of nonlinear system with controller constraints based on finite-approximation-errors ADP algorithm
Ruizhuo Song
,
Wendong Xiao
,
Yanhong Luo
.
adprl 2013
:
19-23
[doi]
Delayed insertion and rule effect moderation of domain knowledge for reinforcement learning
Teck-Hou Teng
,
Ah-Hwee Tan
.
adprl 2013
:
132-139
[doi]
Free energy based policy gradients
Evangelos Theodorou
,
Jiri Najemnik
,
Emanuel Todorov
.
adprl 2013
:
124-131
[doi]
Reinforcement learning in the game of Othello: Learning against a fixed opponent and learning from self-play
M. van der Ree
,
M. Wiering
.
adprl 2013
:
108-115
[doi]
Fault accommodation for complete synchronization of complex neural networks
Zhanshan Wang
,
Fufei Chu
,
Hongjing Liang
,
Huaguang Zhang
.
adprl 2013
:
200-205
[doi]
A novel approach for constructing basis functions in approximate dynamic programming for feedback control
Jian Wang
,
Zhenhua Huang
,
Xin Xu
.
adprl 2013
:
47-51
[doi]
Finite horizon stochastic optimal control of uncertain linear networked control system
Hao Xu
,
Sarangapani Jagannathan
.
adprl 2013
:
24-30
[doi]
Analyzing collective behavior in evolutionary swarm robotic systems based on an ethological approach
Toshiyuki Yasuda
,
Nanami Wada
,
Kazuhiro Ohkura
,
Yoshiyuki Matsumura
.
adprl 2013
:
148-155
[doi]
Finite-horizon optimal control design for uncertain linear discrete-time systems
Qiming Zhao
,
Hao Xu
,
Sarangapani Jagannathan
.
adprl 2013
:
6-12
[doi]
Value function approximation and model predictive control
Mingyuan Zhong
,
M. Johnson
,
Yuval Tassa
,
Tom Erez
,
Emo Todorov
.
adprl 2013
:
100-107
[doi]
2011
2011 IEEE Symposium on Adaptive Dynamic Programming And Reinforcement Learning, ADPRL 2011, Paris, France, April 12-14, 2011
IEEE,
2011.
[doi]
An adaptive-learning framework for semi-cooperative multi-agent coordination
Abdeslem Boukhtouta
,
Jean Berger
,
Warren B. Powell
,
Abraham P. George
.
adprl 2011
:
324-331
[doi]
Path integral control and bounded rationality
Daniel A. Braun
,
Pedro A. Ortega
,
Evangelos Theodorou
,
Stefan Schaal
.
adprl 2011
:
202-209
[doi]
Approximate reinforcement learning: An overview
Lucian Busoniu
,
Damien Ernst
,
Bart De Schutter
,
Robert Babuska
.
adprl 2011
:
1-8
[doi]
Information space receding horizon control
Suman Chakravorty
,
R. Scott Erwin
.
adprl 2011
:
302-309
[doi]
Evolutionary value function approximation
Mohsen Davarynejad
,
Jelmer van Ast
,
Jos L. M. Vrancken
,
Jan van den Berg
.
adprl 2011
:
151-155
[doi]
Near optimal control of mobile robot formations
Travis Dierks
,
Bryan Brenner
,
Sarangapani Jagannathan
.
adprl 2011
:
234-241
[doi]
Higher order Q-Learning
Ashley Edwards
,
William M. Pottenger
.
adprl 2011
:
128-134
[doi]
A reinforcement learning approach for sequential mastery testing
El-Sayed M. El-Alfy
.
adprl 2011
:
295-301
[doi]
Adaptive dynamic programming with balanced weights seeking strategy
Jian Fu
,
Haibo He
,
Zhen Ni
.
adprl 2011
:
210-217
[doi]
Improved neural fitted Q iteration applied to a novel computer gaming and learning benchmark
Thomas Gabel
,
Christian Lutz
,
Martin Riedmiller
.
adprl 2011
:
279-286
[doi]
Safe reinforcement learning in high-risk tasks through policy improvement
Francisco Javier García-Polo
,
Fernando Fernández-Rebollo
.
adprl 2011
:
76-83
[doi]
Model-building semi-Markov adaptive critics
Abhijit Gosavi
,
Susan L. Murray
,
Jiaqiao Hu
.
adprl 2011
:
170-175
[doi]
Structure search of probabilistic models and data correction for EDA-RL
Hisashi Handa
.
adprl 2011
:
332-337
[doi]
Agent self-assessment: Determining policy quality without execution
Alexander Hans
,
Siegmund Duell
,
Steffen Udluft
.
adprl 2011
:
84-90
[doi]
Active exploration for robot parameter selection in episodic reinforcement learning
Oliver Kroemer
,
Jan Peters
.
adprl 2011
:
25-31
[doi]
Higher-level application of Adaptive Dynamic Programming/Reinforcement Learning - a next phase for controls and system identification?
George G. Lendaris
.
adprl 2011
:
[doi]
Application of reinforcement learning-based algorithms in CO2 allowance and electricity markets
Vishnuteja Nanduri
.
adprl 2011
:
164-169
[doi]
Reinforcement learning in multidimensional continuous action spaces
Jason Pazis
,
Michail G. Lagoudakis
.
adprl 2011
:
97-104
[doi]
Dynamic lead time promising
Matthew J. Reindorp
,
Michael C. Fu
.
adprl 2011
:
176-183
[doi]
Directed exploration of policy space using support vector classifiers
Ioannis Rexakis
,
Michail G. Lagoudakis
.
adprl 2011
:
112-119
[doi]
Bayesian active learning with basis functions
Ilya O. Ryzhov
,
Warren B. Powell
.
adprl 2011
:
143-150
[doi]
Complex object manipulation with hierarchical optimal control
Alex Simpkins
,
Emanuel Todorov
.
adprl 2011
:
338-345
[doi]
Grounding subgoals in information transitions
Sander G. van Dijk
,
Daniel Polani
.
adprl 2011
:
105-111
[doi]
Reinforcement learning algorithms for solving classification problems
Marco A. Wiering
,
Hado van Hasselt
,
Auke-Dirk Pietersma
,
Lambert Schomaker
.
adprl 2011
:
91-96
[doi]
Enhancing the episodic natural actor-critic algorithm by a regularisation term to stabilize learning of control structures
Andreas Witsch
,
Roland Reichle
,
Kurt Geihs
,
Sascha Lange
,
Martin Riedmiller
.
adprl 2011
:
156-163
[doi]
Online near optimal control of unknown nonaffine systems with application to HCCI engines
Hassan Zargarzadeh
,
Sarangapani Jagannathan
,
James A. Drallmeier
.
adprl 2011
:
258-263
[doi]
Moving least-squares approximations for linearly-solvable MDP
Mingyuan Zhong
,
Emanuel Todorov
.
adprl 2011
:
218-225
[doi]
Sign in
or
sign up
to see more results.