Abstract is missing.
- Local stability analysis of high-order recurrent neural networks with multi-step piecewise linear activation functionsYujiao Huang, Huaguang Zhang, Dongsheng Yang. 1-5 [doi]
- Finite-horizon optimal control design for uncertain linear discrete-time systemsQiming Zhao, Hao Xu, Sarangapani Jagannathan. 6-12 [doi]
- Adaptive optimal control for nonlinear discrete-time systemsChunbin Qin, Huaguang Zhang, Yanhong Luo. 13-18 [doi]
- Optimal control for a class of nonlinear system with controller constraints based on finite-approximation-errors ADP algorithmRuizhuo Song, Wendong Xiao, Yanhong Luo. 19-23 [doi]
- Finite horizon stochastic optimal control of uncertain linear networked control systemHao Xu, Sarangapani Jagannathan. 24-30 [doi]
- Exponential moving average Q-learning algorithmM. D. Awheda, H. M. Schwartz. 31-38 [doi]
- Real-time tracking on adaptive critic design with uniformly ultimately bounded conditionZhen Ni, Xiao Fang, Haibo He, Dongbin Zhao, Xin Xu. 39-46 [doi]
- A novel approach for constructing basis functions in approximate dynamic programming for feedback controlJian Wang, Zhenhua Huang, Xin Xu. 47-51 [doi]
- A combined hierarchical reinforcement learning based approach for multi-robot cooperative target searching in complex unknown environmentsYifan Cai, Simon X. Yang, Xin Xu. 52-59 [doi]
- The second order temporal difference error for Sarsa(λ)Qi-ming Fu, Quan Liu, Fei Xiao, Guixin Chen. 60-68 [doi]
- Optimistic planning for continuous-action deterministic systemsLucian Busoniu, A. Daniels, Rémi Munos, Robert Babuska. 69-76 [doi]
- Optimistic planning for belief-augmented Markov Decision ProcessesRaphaël Fonteneau, Lucian Busoniu, Rémi Munos. 77-84 [doi]
- Optimized look-ahead trees: Extensions to large and continuous action spacesT. Jung, D. Ernst, F. Maes. 85-92 [doi]
- Bias-corrected Q-learning to control max-operator bias in Q-learningDonghun Lee, Boris Defourny, Warren B. Powell. 93-99 [doi]
- Value function approximation and model predictive controlMingyuan Zhong, M. Johnson, Yuval Tassa, Tom Erez, Emo Todorov. 100-107 [doi]
- Reinforcement learning in the game of Othello: Learning against a fixed opponent and learning from self-playM. van der Ree, M. Wiering. 108-115 [doi]
- A reinforcement learning algorithm developed to model GenCo strategic bidding behavior in multidimensional and continuous state and action spacesA. Y. F. Lau, Dipti Srinivasan, Thomas Reindl. 116-123 [doi]
- Free energy based policy gradientsEvangelos Theodorou, Jiri Najemnik, Emanuel Todorov. 124-131 [doi]
- Delayed insertion and rule effect moderation of domain knowledge for reinforcement learningTeck-Hou Teng, Ah-Hwee Tan. 132-139 [doi]
- Exploring the relationship of reward and punishment in reinforcement learningRobert Lowe, Tom Ziemke. 140-147 [doi]
- Analyzing collective behavior in evolutionary swarm robotic systems based on an ethological approachToshiyuki Yasuda, Nanami Wada, Kazuhiro Ohkura, Yoshiyuki Matsumura. 148-155 [doi]
- Reinforcement learning to train Ms. Pac-Man using higher-order action-relative inputsLuuk Bom, Ruud Henken, Marco Wiering. 156-163 [doi]
- A study on the efficiency of learning a robot controller in various environmentsSachiko Soga, Ichiro Kobayashi. 164-169 [doi]
- On the coordination system for the dimensionality-reduced inputs of marioHisashi Handa. 170-176 [doi]
- Optimal control for a class of nonlinear systems with state delay based on Adaptive Dynamic Programming with ε-error boundXiaofeng Lin, Nuyun Cao, Yuzhang Lin. 177-182 [doi]
- An integrated design for intensified direct heuristic dynamic programmingXiong Luo, Jennie Si, Yuchao Zhou. 183-190 [doi]
- Scalarized multi-objective reinforcement learning: Novel design techniquesKristof Van Moffaert, Madalina M. Drugan, Ann Nowé. 191-199 [doi]
- Fault accommodation for complete synchronization of complex neural networksZhanshan Wang, Fufei Chu, Hongjing Liang, Huaguang Zhang. 200-205 [doi]