The following publications are possibly variants of this publication:
- Pessimistic value iteration for multi-task data sharing in Offline Reinforcement LearningChenjia Bai, Lingxiao Wang 0003, Jianye Hao, Zhuoran Yang, Bin Zhao 0001, Zhen Wang, Xuelong Li 0001. ai, 326:104048, January 2024. [doi]
- Pessimistic Bootstrapping for Uncertainty-Driven Offline Reinforcement LearningChenjia Bai, Lingxiao Wang 0003, Zhuoran Yang, Zhi-Hong Deng, Animesh Garg, Peng Liu 0008, Zhaoran Wang. iclr 2022: [doi]
- Pessimistic Model Selection for Offline Deep Reinforcement LearningChao-Han Huck Yang, Zhengling Qi, Yifan Cui 0001, Pin-Yu Chen. uai 2023: 2379-2389 [doi]
- Neural Network Approximation for Pessimistic Offline Reinforcement LearningDi Wu, Yuling Jiao, Li Shen, Haizhao Yang, Xiliang Lu. AAAI 2024: 15868-15877 [doi]
- Pessimistic Minimax Value Iteration: Provably Efficient Equilibrium Learning from Offline DatasetsHan Zhong, Wei Xiong, Jiyuan Tan, Liwei Wang, Tong Zhang, Zhaoran Wang, Zhuoran Yang. icml 2022: 27117-27142 [doi]
- Personalized robotic control via constrained multi-objective reinforcement learningXiangkun He, Zhongxu Hu, Haohan Yang, Chen Lv. ijon, 565:126986, January 2024. [doi]