The following publications are possibly variants of this publication:
- Conservative Offline Distributional Reinforcement LearningYecheng Jason Ma, Dinesh Jayaraman, Osbert Bastani. nips 2021: 19235-19247 [doi]
- Mildly Conservative Q-Learning for Offline Reinforcement LearningJiafei Lyu, Xiaoteng Ma, Xiu Li, Zongqing Lu. nips 2022: [doi]
- Conservative State Value Estimation for Offline Reinforcement LearningLiting Chen, Jie Yan, Zhengdao Shao, Lu Wang, Qingwei Lin, Saravanakumar Rajmohan, Thomas Moscibroda, Dongmei Zhang. nips 2023: [doi]
- Offline Quantum Reinforcement Learning in a Conservative MannerZhihao Cheng, Kaining Zhang, Li Shen 0008, Dacheng Tao. AAAI 2023: 7148-7156 [doi]
- OCEAN-MBRL: Offline Conservative Exploration for Model-Based Offline Reinforcement LearningFan Wu, Rui Zhang, Qi Yi, Yunkai Gao, Jiaming Guo, Shaohui Peng, Siming Lan, Husheng Han, Yansong Pan, Kaizhao Yuan, Pengwei Jin, Ruizhi Chen, Yunji Chen, Ling Li. AAAI 2024: 15897-15905 [doi]
- Corruption-robust Offline Reinforcement LearningXuezhou Zhang, Yiding Chen, Xiaojin Zhu 0001, Wen Sun. aistats 2022: 5757-5773 [doi]
- CLARE: Conservative Model-Based Reward Learning for Offline Inverse Reinforcement LearningSheng Yue, Guanbo Wang, Wei Shao 0006, Zhaofeng Zhang, Sen Lin, Ju Ren, Junshan Zhang. iclr 2023: [doi]
- Counterfactual Conservative Q Learning for Offline Multi-agent Reinforcement LearningJianzhun Shao, Yun Qu, Chen Chen, Hongchang Zhang, Xiangyang Ji. nips 2023: [doi]