Abstract is missing.
- Multiagent Learning: From Fundamentals to Foundation ModelsKarl Tuyls. 1 [doi]
- Common Sense: The Dark Matter of Language and IntelligenceYejin Choi. 2 [doi]
- Geometric Principles of Individual and Collective Decision-MakingIain D. Couzin. 3 [doi]
- Proportionality in Multiwinner Voting: The Power of Local SearchEdith Elkind. 4 [doi]
- Trust Region Bounds for Decentralized PPO Under Non-stationarityMingfei Sun, Sam Devlin, Jacob Beck, Katja Hofmann, Shimon Whiteson. 5-13 [doi]
- Multi-Agent Reinforcement Learning for Adaptive Mesh RefinementJiachen Yang, Ketan Mittal, Tarik Dzanic, Socratis Petrides, Brendan Keith, Brenden K. Petersen, Daniel M. Faissol, Robert W. Anderson. 14-22 [doi]
- Adaptive Learning Rates for Multi-Agent Reinforcement LearningJiechuan Jiang, Zongqing Lu. 23-30 [doi]
- Adaptive Value Decomposition with Greedy Marginal Contribution Computation for Cooperative Multi-Agent Reinforcement LearningShanqi Liu, Yujing Hu, Runze Wu, Dong Xing, Yu Xiong, Changjie Fan, Kun Kuang, Yong Liu. 31-39 [doi]
- A Variational Approach to Mutual Information-Based Coordination for Multi-Agent Reinforcement LearningWoojun Kim, Whiyoung Jung, Myungsik Cho, Youngchul Sung. 40-48 [doi]
- Mediated Multi-Agent Reinforcement LearningDmitry Ivanov, Ilya Zisman, Kirill Chernyshev. 49-57 [doi]
- EXPODE: EXploiting POlicy Discrepancy for Efficient Exploration in Multi-agent Reinforcement LearningYucong Zhang, Chao Yu. 58-66 [doi]
- TiZero: Mastering Multi-Agent Football with Curriculum Learning and Self-PlayFanqi Lin, Shiyu Huang, Tim Pearce, Wenze Chen, Wei-Wei Tu. 67-76 [doi]
- Ask and You Shall be Served: Representing & Solving Multi-agentOptimization Problems with Service Requesters and ProvidersMaya Lavie, Tehila Caspi, Omer Lev, Roie Zivan. 77-85 [doi]
- Fairness Driven Efficient Algorithms for Sequenced Group Trip Planning Query ProblemNapendra Solanki, Shweta Jain 0002, Suman Banerjee 0002, Yayathi Pavan Kumar S. 86-94 [doi]
- Domain-Independent Deceptive PlanningAdrian Price, Ramon Fraga Pereira, Peta Masters, Mor Vered. 95-103 [doi]
- CAMS: Collision Avoiding Max-Sum for Mobile Sensor TeamsArseni Pertzovskiy, Roie Zivan, Noa Agmon. 104-112 [doi]
- Risk-Constrained Planning for Multi-Agent Systems with Shared ResourcesAnna Gautier, Marc Rigter, Bruno Lacerda, Nick Hawes, Michael J. Wooldridge. 113-121 [doi]
- Quantitative Planning with Action Deception in Concurrent Stochastic GamesChongyang Shi, Shuo Han 0002, Jie Fu. 122-130 [doi]
- Towards Computationally Efficient Responsibility Attribution in Decentralized Partially Observable MDPsStelios Triantafyllou, Goran Radanovic. 131-139 [doi]
- On-line Estimators for Ad-hoc Task Execution: Learning Types and Parameters of Teammates for Effective TeamworkMatheus Aparecido do Carmo Alves, Elnaz Shafipour Yourdshahi, Amokh Varma, Leandro Soriano Marcolino, Jó Ueyama, Plamen P. Angelov. 140-142 [doi]
- Fair Allocation of Two Types of ChoresHaris Aziz 0001, Jeremy Lindsay, Angus Ritossa, Mashbat Suzuki. 143-151 [doi]
- Fairly Dividing Mixtures of Goods and Chores under Lexicographic PreferencesHadi Hosseini, Sujoy Sikdar, Rohit Vaish, Lirong Xia. 152-160 [doi]
- Graphical House AllocationHadi Hosseini, Justin Payan, Rik Sengupta, Rohit Vaish, Vignesh Viswanathan. 161-169 [doi]
- Approximation Algorithm for Computing Budget-Feasible EF1 AllocationsJiarui Gan, Bo Li 0037, Xiaowei Wu 0001. 170-178 [doi]
- Yankee Swap: A Fast and Simple Fair Allocation Mechanism for Matroid Rank ValuationsVignesh Viswanathan, Yair Zick. 179-187 [doi]
- Fairness in the Assignment Problem with Uncertain PrioritiesZeyu Shen 0001, Zhiyi Wang, Xingyu Zhu 0003, Brandon Fain, Kamesh Munagala. 188-196 [doi]
- Possible Fairness for Allocating Indivisible ResourcesHaris Aziz 0001, Bo Li, Shiji Xing, Yu Zhou. 197-205 [doi]
- Efficient Nearly-Fair Division with Capacity ConstraintsHila Shoshan, Noam Hazon, Erel Segal-haLevi. 206-214 [doi]
- Equilibria and Convergence in Fire Sale GamesNils Bertschinger, Martin Hoefer 0001, Simon Krogmann, Pascal Lenzner, Steffen Schuldenzucker, Lisa Wilhelmi. 215-223 [doi]
- Bridging the Gap Between Single and Multi Objective GamesWillem Röpke, Carla Groenland, Roxana Radulescu, Ann Nowé, Diederik M. Roijers. 224-232 [doi]
- Is Nash Equilibrium Approximator Learnable?Zhijian Duan 0001, Wenhan Huang, Dinghuai Zhang, Yali Du 0001, Jun Wang, Yaodong Yang 0001, Xiaotie Deng. 233-241 [doi]
- Learning the Stackelberg Equilibrium in a Newsvendor GameNicolò Cesa-Bianchi, Tommaso Cesari, Takayuki Osogami, Marco Scarsini, Segev Wasserkrug. 242-250 [doi]
- Hedonic Games With Friends, Enemies, and Neutrals: Resolving Open Questions and Fine-Grained ComplexityJiehua Chen 0001, Gergely Csáji, Sanjukta Roy, Sofia Simola. 251-259 [doi]
- Debt Transfers in Financial Networks: Complexity and EquilibriaPanagiotis Kanellopoulos, Maria Kyropoulou, Hao Zhou. 260-268 [doi]
- A Study of Nash Equilibria in Multi-Objective Normal-Form GamesWillem Röpke, Diederik M. Roijers, Ann Nowé, Roxana Radulescu. 269-271 [doi]
- Learning Properties in Simulation-Based GamesCyrus Cousins, Bhaskar Mishra, Enrique Areyan Viqueira, Amy Greenwald. 272-280 [doi]
- Establishing Shared Query Understanding in an Open Multi-Agent SystemNikolaos Kondylidis, Ilaria Tiddi, Annette ten Teije. 281-289 [doi]
- Communicating Agent Intentions for Human-Agent Decision Making under UncertaintyJulie Porteous, Alan Lindsay, Fred Charles. 290-298 [doi]
- Trusting Artificial Agents: Communication Trumps PerformanceMarin Le Guillou, Laurent Prévot 0001, Bruno Berberian. 299-306 [doi]
- Nonverbal Human Signals Can Help Autonomous Agents Infer Human Preferences for Their BehaviorKate Candon, Jesse Chen, Yoony Kim, Zoe Hsu, Nathan Tsoi, Marynel Vázquez. 307-316 [doi]
- On Subset Selection of Multiple Humans To Improve Human-AI Team AccuracySagalpreet Singh, Shweta Jain 0002, Shashi Shekhar Jha. 317-325 [doi]
- Do Explanations Improve the Quality of AI-assisted Human Decisions? An Algorithm-in-the-Loop Analysis of Factual & Counterfactual ExplanationsLujain Ibrahim, Mohammad M. Ghassemi, Tuka Alhanai. 326-334 [doi]
- Automated Task-Time Interventions to Improve Teamwork using Imitation LearningSangwon Seo, Bing Han, Vaibhav V. Unhelkar. 335-344 [doi]
- Should My Agent Lie for Me? A Study on Attitudes of US-basedParticipants Towards Deceptive AI in Selected Future-of-workStefan Sarkadi, Peidong Mei, Edmond Awad. 345-354 [doi]
- A Logic of Only-Believing over Arbitrary Probability DistributionsQihui Feng, Daxin Liu 0002, Vaishak Belle, Gerhard Lakemeyer. 355-363 [doi]
- A Deontic Logic of Knowingly ComplyingCarlos Areces, Valentin Cassano, Pablo F. Castro, Raul Fervari, Andrés R. Saravia. 364-372 [doi]
- Learning Logic Specifications for Soft Policy Guidance in POMCPGiulio Mazzi, Daniele Meli, Alberto Castellini, Alessandro Farinelli. 373-381 [doi]
- Strategic (Timed) Computation Tree LogicJaime Arias 0001, Wojciech Jamroga, Wojciech Penczek, Laure Petrucci, Teofil Sidoruk. 382-390 [doi]
- Attention! Dynamic Epistemic Logic Models of (In)attentive AgentsGaia Belardinelli, Thomas Bolander. 391-399 [doi]
- (Arbitrary) Partial CommunicationRustam Galimullin, Fernando R. Velázquez-Quesada. 400-408 [doi]
- Epistemic Abstract Argumentation Framework: Formal Foundations, Computation and ComplexityGianvincenzo Alfano, Sergio Greco, Francesco Parisi, Irina Trubitsyna. 409-417 [doi]
- Actions, Continuous Distributions and Meta-BeliefsVaishak Belle. 418-426 [doi]
- AC2C: Adaptively Controlled Two-Hop Communication for Multi-Agent Reinforcement LearningXuefeng Wang, Xinran Li, Jiawei Shao, Jun Zhang 0004. 427-435 [doi]
- Learning Structured Communication for Multi-Agent Reinforcement LearningJunjie Sheng, Xiangfeng Wang, Bo Jin 0003, Wenhao Li, Jun Wang, Junchi Yan, Tsung-Hui Chang, Hongyuan Zha. 436-438 [doi]
- Model-based Sparse Communication in Multi-agent Reinforcement LearningShuai Han, Mehdi Dastani, Shihan Wang. 439-447 [doi]
- Get It in Writing: Formal Contracts Mitigate Social Dilemmas in Multi-Agent RLPhillip J. K. Christoffersen, Andreas A. Haupt, Dylan Hadfield-Menell. 448-456 [doi]
- The Benefits of Power Regularization in Cooperative Reinforcement LearningMichelle Li, Michael Dennis 0001. 457-465 [doi]
- MAC-PO: Multi-Agent Experience Replay via Collective Priority OptimizationYongsheng Mei, Hanhan Zhou, Tian Lan, Guru Venkataramani, Peng Wei. 466-475 [doi]
- Self-Motivated Multi-Agent ExplorationShaowei Zhang, Jiahan Cao, Lei Yuan, Yang Yu 0001, De-Chuan Zhan. 476-484 [doi]
- Sequential Cooperative Multi-Agent Reinforcement LearningYifan Zang, Jinmin He, Kai Li, Haobo Fu, Qiang Fu, Junliang Xing. 485-493 [doi]
- Online Coalitional Skill FormationSaar Cohen 0001, Noa Agmon. 494-503 [doi]
- Multi-Agent Consensus-based Bundle Allocation for Multi-mode Composite TasksGauthier Picard. 504-512 [doi]
- Allocation Problem in Remote Teleoperation: Online Matching with Offline Reusable Resources and Delayed AssignmentsOsnat Ackerman Viden, Yohai Trabelsi, Pan Xu 0001, Karthik Abinav Sankararaman, Oleg Maksimov, Sarit Kraus. 513-521 [doi]
- Optimal Coalition Structures for Probabilistically Monotone Partition Function GamesShaheen Fatima, Michael J. Wooldridge. 522-524 [doi]
- A Comparison of New Swarm Task Allocation Algorithms in Unknown Environments with Varying Task DensityGrace Cai, Noble Harasha, Nancy A. Lynch. 525-533 [doi]
- Abstracting Noisy Robot ProgramsTill Hofmann, Vaishak Belle. 534-542 [doi]
- Structural Credit Assignment-Guided Coordinated MCTS: An Efficient and Scalable Method for Online Multiagent PlanningQian Che, Wanyuan Wang, Fengchen Wang, Tianchi Qiao, Xiang Liu, Jiuchuan Jiang, Bo An 0001, Yichuan Jiang. 543-551 [doi]
- Strategic Planning for Flexible Agent Availability in Large Taxi FleetsRajiv Ranjan Kumar, Pradeep Varakantham, Shih-Fen Cheng. 552-560 [doi]
- Equitability and Welfare Maximization for Allocating Indivisible ItemsAnkang Sun, Bo Chen 0002, Xuan Vinh Doan. 561-563 [doi]
- Best of Both Worlds: Agents with EntitlementsMartin Hoefer 0001, Marco Schmalhofer, Giovanna Varricchio. 564-572 [doi]
- Mitigating Skewed Bidding for Conference Paper AssignmentInbal Rozencweig, Reshef Meir, Nicholas Mattei, Ofra Amir. 573-581 [doi]
- Price of Anarchy in a Double-Sided Critical Distribution SystemDavid Sychrovsky, Jakub Cerný, Sylvain Lichau, Martin Loebl. 582-590 [doi]
- Improved EFX Approximation Guarantees under Ordinal-based AssumptionsEvangelos Markakis, Christodoulos Santorinaios. 591-599 [doi]
- Assigning Agents to Increase Network-Based Neighborhood DiversityZirou Qiu, Andrew Yuan, Chen Chen, Madhav V. Marathe, S. S. Ravi, Daniel J. Rosenkrantz, Richard Edwin Stearns, Anil Vullikanti. 600-608 [doi]
- Altruism, Collectivism and Egalitarianism: On a Variety of Prosocial Behaviors in Binary Networked Public Goods GamesJichen Li, Xiaotie Deng, Yukun Cheng, Yuqi Pan, Xuanzhi Xia, Zongjun Yang, Jan Xie. 609-624 [doi]
- The Role of Space, Density and Migration in Social DilemmasJacques Bara, Fernando P. Santos, Paolo Turrini. 625-633 [doi]
- Non-strategic Econometrics (for Initial Play)Daniel Chui, Jason D. Hartline, James R. Wright. 634-642 [doi]
- Efficient Stackelberg Strategies for Finitely Repeated GamesNatalie Collina, Eshwar Ram Arunachaleswaran, Michael Kearns. 643-651 [doi]
- Learning Density-Based Correlated Equilibria for Markov GamesLibo Zhang, Yang Chen, Toru Takisaka, Bakh Khoussainov, Michael Witbrock, Jiamou Liu. 652-660 [doi]
- IRS: An Incentive-compatible Reward Scheme for AlgorandMaizi Liao, Wojciech Golab, Seyed Majid Zahedi. 661-669 [doi]
- Data Structures for Deviation PayoffsBryce Wiedenbeck, Erik Brinkman. 670-678 [doi]
- PECAN: Leveraging Policy Ensemble for Context-Aware Zero-Shot Human-AI CoordinationXingzhou Lou, Jiaxian Guo, Junge Zhang, Jun Wang 0012, Kaiqi Huang, Yali Du 0001. 679-688 [doi]
- Semi-Autonomous Systems with Contextual Competence AwarenessSaaduddin Mahmud, Connor Basich, Shlomo Zilberstein. 689-697 [doi]
- Joint Engagement Classification using Video Augmentation Techniques for Multi-person HRI in the wildYubin Kim, Huili Chen, Sharifa Alghowinem, Cynthia Breazeal, Hae Won Park. 698-707 [doi]
- Multiagent Inverse Reinforcement Learning via Theory of Mind ReasoningHaochen Wu, Pedro Sequeira, David V. Pynadath. 708-716 [doi]
- Persuading to Prepare for Quitting Smoking with a Virtual Coach: Using States and User Characteristics to Predict BehaviorNele Albers, Mark A. Neerincx, Willem-Paul Brinkman. 717-726 [doi]
- Think Twice: A Human-like Two-stage Conversational Agent for Emotional Response GenerationYushan Qian, Bo Wang, Shangzhao Ma, Bin Wu, Shuo Zhang, Dongming Zhao, Kun Huang, Yuexian Hou. 727-736 [doi]
- Generating Stylistic and Personalized Dialogues for Virtual Agents in NarrativesWeilai Xu, Fred Charles, Charlie Hargood. 737-746 [doi]
- Reducing Racial Bias by Interacting with Virtual Agents: An Intervention in Virtual RealityDavid Obremski, Ohenewa Bediako Akuffo, Leonie Lücke, Miriam Semineth, Sarah Tomiczek, Hanna-Finja Weichert, Birgit Lugrin. 747-755 [doi]
- Provable Optimization of Quantal Response Leader-Follower Games with Exponentially Large Action SpacesJinzhao Li, Daniel Fink 0002, Christopher Wood, Carla P. Gomes, Yexiang Xue. 756-765 [doi]
- Playing to Learn, or to Keep Secret: Alternating-Time Logic Meets Information TheoryMasoud Tabatabaei, Wojciech Jamroga. 766-774 [doi]
- Synthesis of Resource-Aware Controllers Against Rational AgentsRodica Condurache, Catalin Dima, Youssouf Oualhadj, Nicolas Troquard. 775-783 [doi]
- Computationally Feasible StrategiesCatalin Dima, Wojciech Jamroga. 784-792 [doi]
- Towards the Verification of Strategic Properties in Multi-Agent Systems with Imperfect InformationAngelo Ferrando 0001, Vadim Malvone. 793-801 [doi]
- Follow your Nose: Using General Value Functions for Directed Exploration in Reinforcement LearningDurgesh Kalwar, Omkar Shelke, Somjit Nath, Hardik Meisheri, Harshad Khadilkar. 802-809 [doi]
- FedFormer: Contextual Federation with Attention in Reinforcement LearningLiam Hebert, Lukasz Golab, Pascal Poupart, Robin Cohen. 810-818 [doi]
- Diverse Policy Optimization for Structured Action SpaceWenhao Li, Baoxiang Wang 0001, Shanchao Yang, Hongyuan Zha. 819-828 [doi]
- Enhancing Reinforcement Learning Agents with Local GuidesPaul Daoudi, Bogdan Robu, Christophe Prieur 0001, Ludovic Dos Santos, Merwan Barlier. 829-838 [doi]
- Scalar Reward is Not EnoughPeter Vamplew 0001, Benjamin J. Smith, Johan Källström, Gabriel de Oliveira Ramos, Roxana Radulescu, Diederik M. Roijers, Conor F. Hayes, Friedrik Hentz, Patrick Mannion, Pieter J. K. Libin, Richard Dazeley, Cameron Foale. 839-841 [doi]
- Targeted Search Control in AlphaZero for Effective Policy ImprovementAlexandre Trudeau, Michael Bowling. 842-850 [doi]
- Out-of-Distribution Detection for Reinforcement Learning Agents with Probabilistic Dynamics ModelsTom Haider, Karsten Roscher, Felippe Schmoeller da Roza, Stephan Günnemann. 851-859 [doi]
- Knowledge Compilation for Constrained Combinatorial Action Spaces in Reinforcement LearningJiajing Ling, Moritz Lukas Schuler, Akshat Kumar, Pradeep Varakantham. 860-868 [doi]
- Anonymous Multi-Agent Path Finding with Individual DeadlinesGilad Fine, Dor Atzmon, Noa Agmon. 869-877 [doi]
- Learn to Solve the Min-max Multiple Traveling Salesmen Problem with Reinforcement LearningJunyoung Park, Changhyun Kwon, Jinkyoo Park. 878-886 [doi]
- Counterfactual Fairness Filter for Fair-Delay Multi-Robot NavigationHikaru Asano, Ryo Yonetani, Mai Nishimura, Tadashi Kozuno. 887-895 [doi]
- Improved Complexity Results and an Efficient Solution for Connected Multi-Agent Path FindingIsseïnie Calviac, Ocan Sankur, François Schwarzentruber. 896-904 [doi]
- Optimally Solving the Multiple Watchman Route Problem with Heuristic SearchYaakov Livne, Dor Atzmon, Shawn Skyler, Eli Boyarski, Amir Shapiro, Ariel Felner. 905-913 [doi]
- Distributed Planning with Asynchronous Execution with Local Navigation for Multi-agent Pickup and Delivery ProblemYuki Miyashita, Tomoki Yamauchi, Toshiharu Sugawara. 914-922 [doi]
- Energy-aware UAV Path Planning with Adaptive SpeedJonathan Diller, Qi Han. 923-931 [doi]
- Coordination of Multiple Robots along Given Paths with Bounded Junction ComplexityMikkel Abrahamsen, Tzvika Geft, Dan Halperin, Barak Ugav. 932-940 [doi]
- Best of Both Worlds Fairness under EntitlementsHaris Aziz 0001, Aditya Ganguly, Evi Micha. 941-948 [doi]
- Probabilistic Rationing with Categorized Priorities: Processing Reserves Fairly and EfficientlyHaris Aziz 0001. 949-956 [doi]
- Semi-Popular Matchings and Copeland WinnersTelikepalli Kavitha, Rohit Vaish. 957-965 [doi]
- Host Community Respecting Refugee HousingDusan Knop, Simon Schierreich. 966-975 [doi]
- Online Matching with Delays and Stochastic Arrival TimesMathieu Mari, Michal Pawlowski, Runtian Ren, Piotr Sankowski. 976-984 [doi]
- Adapting Stable Matchings to Forced and Forbidden PairsNiclas Boehmer, Klaus Heeger. 985-993 [doi]
- Stable Marriage in Euclidean SpaceYinghui Wen, Zhongyi Zhang, Jiong Guo. 994-1002 [doi]
- A Map of Diverse Synthetic Stable Roommates InstancesNiclas Boehmer, Klaus Heeger, Stanislaw Szufa. 1003-1011 [doi]
- Empirical Game-Theoretic Analysis for Mean Field GamesYongzhao Wang, Michael P. Wellman. 1025-1033 [doi]
- Differentiable Arbitrating in Zero-sum Markov GamesJing Wang, Meichen Song, Feng Gao, Boyi Liu, Zhaoran Wang, Yi Wu. 1034-1043 [doi]
- Learning Parameterized Families of GamesMadelyn Gatchel, Bryce Wiedenbeck. 1044-1052 [doi]
- Fictitious Cross-Play: Learning Global Nash Equilibrium in Mixed Cooperative-Competitive GamesZelai Xu, Yancheng Liang, Chao Yu 0005, Yu Wang, Yi Wu. 1053-1061 [doi]
- Cost Inference for Feedback Dynamic Games from Noisy Partial State Observations and Incomplete TrajectoriesJingqi Li, Chih-Yuan Chiu, Lasse Peters, Somayeh Sojoudi, Claire J. Tomlin, David Fridovich-Keil. 1062-1070 [doi]
- Multiplicative Weight Updates for Extensive Form GamesChirag Chhablani, Michael Sullins, Ian A. Kash. 1071-1078 [doi]
- A Hybrid Framework of Reinforcement Learning and Physics-Informed Deep Learning for Spatiotemporal Mean Field GamesXu Chen, Shuo Liu, Xuan Di. 1079-1087 [doi]
- Adversarial Inverse Reinforcement Learning for Mean Field GamesYang Chen, Libo Zhang, Jiamou Liu, Michael Witbrock. 1088-1096 [doi]
- GANterfactual-RL: Understanding Reinforcement Learning Agents' Strategies through Visual Counterfactual ExplanationsTobias Huber, Maximilian Demmler, Silvan Mertes, Matthew L. Olson, Elisabeth André. 1097-1106 [doi]
- Asynchronous Multi-Agent Reinforcement Learning for Efficient Real-Time Multi-Robot Cooperative ExplorationChao Yu, Xinyi Yang, Jiaxuan Gao, Jiayu Chen, Yunfei Li, Jijia Liu, Yunfei Xiang, Ruixin Huang, Huazhong Yang, Yi Wu, Yu Wang. 1107-1115 [doi]
- Dec-AIRL: Decentralized Adversarial IRL for Human-Robot TeamingPrasanth Sengadu Suresh, Yikang Gui, Prashant Doshi. 1116-1124 [doi]
- Structural Attention-based Recurrent Variational Autoencoder for Highway Vehicle Anomaly DetectionNeeloy Chakraborty, Aamir Hasan 0002, Shuijing Liu, Tianchen Ji, Weihang Liang, D. Livingston McPherson, Katherine Rose Driggs-Campbell. 1125-1134 [doi]
- Controlled Diversity with Preference : Towards Learning a Diverse Set of Desired SkillsMaxence Hussonnois, Thommen George Karimpanal, Santu Rana. 1135-1143 [doi]
- Learning from Multiple Independent Advisors in Multi-agent Reinforcement LearningSriram Ganapathi Subramanian, Matthew E. Taylor, Kate Larson, Mark Crowley 0001. 1144-1153 [doi]
- Kiko: Programming Agents to Enact Interaction ModelsSamuel H. Christie V., Munindar P. Singh, Amit K. Chopra. 1154-1163 [doi]
- CraftEnv: A Flexible Collective Robotic Construction Environment for Multi-Agent Reinforcement LearningRui Zhao, Xu Liu, Yizheng Zhang, Minghao Li, Cheng Zhou, Shuai Li, Lei Han. 1164-1172 [doi]
- Feedback-Guided Intention Scheduling for BDI AgentsMichael Dann, John Thangarajah, Minyi Li 0001. 1173-1181 [doi]
- A Behaviour-Driven Approach for Testing Requirements via User and System Stories in Agent SystemsSebastian Rodriguez, John Thangarajah, Michael Winikoff. 1182-1190 [doi]
- ML-MAS: A Hybrid AI Framework for Self-Driving VehiclesHilal Al Shukairi, Rafael C. Cardoso 0001. 1191-1199 [doi]
- Signifiers as a First-class Abstraction in Hypermedia Multi-Agent SystemsDanai Vachtsevanou, Andrei Ciortea, Simon Mayer, Jérémy Lemée. 1200-1208 [doi]
- MAIDS - A Framework for the Development of Multi-Agent Intentional Dialogue SystemsDébora C. Engelmann, Alison R. Panisson, Renata Vieira, Jomi Fred Hübner, Viviana Mascardi, Rafael H. Bordini. 1209-1217 [doi]
- Mandrake: Multiagent Systems as a Basis for Programming Fault-Tolerant Decentralized ApplicationsSamuel H. Christie V., Munindar P. Singh, Amit K. Chopra. 1218-1220 [doi]
- Curriculum Offline Reinforcement LearningYuanying Cai, Chuheng Zhang, Hanye Zhao, Li Zhao 0007, Jiang Bian 0002. 1221-1229 [doi]
- Decentralized Model-Free Reinforcement Learning in Stochastic Games with Average-Reward ObjectiveRomain Cravic, Nicolas Gast, Bruno Gaujal. 1230-1238 [doi]
- Less Is More: Refining Datasets for Offline Reinforcement Learning with Reward MachinesHaoyuan Sun, Feng Wu 0001. 1239-1247 [doi]
- A Self-Organizing Neuro-Fuzzy Q-Network: Systematic Design with Offline Hybrid LearningJohn Wesley Hostetter, Mark Abdelshiheed, Tiffany Barnes, Min Chi. 1248-1257 [doi]
- Learning to Coordinate from Offline Datasets with Uncoordinated Behavior PoliciesJinming Ma, Feng Wu. 1258-1266 [doi]
- D-Shape: Demonstration-Shaped Reinforcement Learning via Goal-ConditioningCaroline Wang, Garrett Warnell, Peter Stone. 1267-1275 [doi]
- How To Guide Your Learner: Imitation Learning with Active Adaptive Expert InvolvementXu-Hui Liu, Feng Xu, Xinyu Zhang, Tianyuan Liu, Shengyi Jiang, Ruifeng Chen 0003, Zongzhang Zhang, Yang Yu 0001. 1276-1284 [doi]
- Imitating Opponent to Win: Adversarial Policy Imitation Learning in Two-player Competitive GamesThe Viet Bui, Tien Mai, Thanh Hong Nguyen. 1285-1293 [doi]
- Indexability is Not Enough for Whittle: Improved, Near-Optimal Algorithms for Restless BanditsAbheek Ghosh, Dheeraj Nagaraj, Manish Jain, Milind Tambe. 1294-1302 [doi]
- Avoiding Starvation of Arms in Restless Multi-Armed BanditsDexun Li, Pradeep Varakantham. 1303-1311 [doi]
- Restless Multi-Armed Bandits for Maternal and Child Health: Results from Decision-Focused LearningShresth Verma, Aditya Mate, Kai Wang 0040, Neha Madhiwalla, Aparna Hegde, Aparna Taneja, Milind Tambe. 1312-1320 [doi]
- Fairness for Workers Who Pull the Arms: An Index Based Policy for Allocation of Restless Bandit TasksArpita Biswas, Jackson A. Killian, Paula Rodriguez Diaz, Susobhan Ghosh, Milind Tambe. 1321-1328 [doi]
- On Regret-optimal Cooperative Nonstochastic Multi-armed BanditsJialin Yi, Milan Vojnovic. 1329-1335 [doi]
- Equilibrium Bandits: Learning Optimal Equilibria of Unknown DynamicsSiddharth Chandak, Ilai Bistritz, Nicholas Bambos. 1336-1344 [doi]
- ExPoSe: Combining State-Based Exploration with Gradient-Based Online SearchDixant Mittal, Siddharth Aravindan, Wee Sun Lee. 1345-1353 [doi]
- Formally-Sharp DAgger for MCTS: Lower-Latency Monte Carlo Tree Search using Data Aggregation with Formal MethodsDebraj Chakraborty 0002, Damien Busatto-Gaston, Jean-François Raskin, Guillermo A. Pérez 0001. 1354-1362 [doi]
- Price of Anarchy for First Price Auction with Risk-Averse BiddersZhiqiang Zhuang, Kewen Wang 0001, Zhe Wang 0001. 1363-1369 [doi]
- A Redistribution Framework for Diffusion AuctionsSizhe Gu, Yao Zhang, Yida Zhao, Dengji Zhao. 1370-1378 [doi]
- Sybil-Proof Diffusion Auction in Social NetworksHongyin Chen, Xiaotie Deng, Ying Wang, Yue Wu, Dengji Zhao. 1379-1387 [doi]
- Representing and Reasoning about AuctionsMunyque Mittelmann, Laurent Perrussel, Sylvain Bouveret. 1388-1390 [doi]
- Revisiting the Distortion of Distributed VotingAris Filos-Ratsikas, Alexandros A. Voudouris. 1391-1399 [doi]
- Bounded Approval Ballots: Balancing Expressiveness and Simplicity for Multiwinner ElectionsDorothea Baumeister, Linus Boes, Christian Laußmann, Simon Rey. 1400-1408 [doi]
- On the Distortion of Single Winner Elections with Aligned CandidatesDimitris Fotakis 0001, Laurent Gourvès. 1409-1411 [doi]
- SAT-based Judgment AggregationAri Conati, Andreas Niskanen, Matti Järvisalo. 1412-1420 [doi]
- Decentralised and Cooperative Control of Multi-Robot Systems through Distributed OptimisationYi Dong, Zhongguo Li, Xingyu Zhao 0001, Zhengtao Ding, Xiaowei Huang 0001. 1421-1429 [doi]
- Byzantine Resilience at Swarm Scale: A Decentralized Blocklist Protocol from Inter-robot AccusationsKacper Wardega, Max von Hippel, Roberto Tron, Cristina Nita-Rotaru, Wenchao Li 0001. 1430-1438 [doi]
- Stigmergy-based, Dual-Layer Coverage of Unknown RegionsOri Rappel, Michael Amir, Alfred M. Bruckstein. 1439-1447 [doi]
- Mitigating Imminent Collision for Multi-robot Navigation: A TTC-force Reward Shaping ApproachJinlin Chen, Jiannong Cao 0001, Zhiqin Cheng, Wei Li 0121. 1448-1456 [doi]
- Gathering of Anonymous AgentsArnhav Datar, Nischith Shadagopan M. N, John Augustine. 1457-1465 [doi]
- Safe Deep Reinforcement Learning by Verifying Task-Level PropertiesEnrico Marchesini, Luca Marzari, Alessandro Farinelli, Christopher Amato. 1466-1475 [doi]
- Decentralized Safe Navigation for Multi-agent Systems via Risk-aware Weighted Buffered Voronoi CellsYiwei Lyu, John M. Dolan, Wenhao Luo. 1476-1484 [doi]
- Heterogeneous Multi-Robot Reinforcement LearningMatteo Bettini, Ajay Shankar, Amanda Prorok. 1485-1494 [doi]
- Efficient Interactive Recommendation via Huffman Tree-based Policy LearningLongxiang Shi, Zilin Zhang, Shoujin Wang, Binbin Zhou, Minghui Wu, Cheng Yang, Shijian Li. 1495-1503 [doi]
- HOPE: Human-Centric Off-Policy Evaluation for E-Learning and HealthcareGe Gao, Song Ju, Markel Sanz Ausin, Min Chi. 1504-1513 [doi]
- ShelfHelp: Empowering Humans to Perform Vision-Independent Manipulation Tasks with a Socially Assistive Robotic CaneShivendra Agrawal, Suresh Nayak, Ashutosh Naik, Bradley Hayes. 1514-1523 [doi]
- Preference-Aware Delivery Planning for Last-Mile LogisticsQian Shao, Shih-Fen Cheng. 1524-1532 [doi]
- Multi-Agent Reinforcement Learning with Safety Layer for Active Voltage ControlYufeng Shi, Mingxiao Feng, Minrui Wang, Wengang Zhou, Houqiang Li. 1533-1541 [doi]
- Multi-agent Signalless Intersection Management with Dynamic Platoon FormationPhuriwat Worrawichaipat, Enrico H. Gerding, Ioannis Kaparias, Sarvapali D. Ramchurn. 1542-1550 [doi]
- SocialLight: Distributed Cooperation Learning towards Network-Wide Traffic Signal ControlHarsh Goel, Yifeng Zhang, Mehul Damani, Guillaume Sartoretti. 1551-1559 [doi]
- Model-Based Reinforcement Learning for Auto-bidding in Display AdvertisingShuang Chen, Qisen Xu, Liang Zhang, Yongbo Jin, Wenhao Li, Linjian Mo. 1560-1568 [doi]
- Learning Inter-Agent Synergies in Asymmetric Multiagent SystemsGaurav Dixit, Kagan Tumer. 1569-1577 [doi]
- Asymptotic Convergence and Performance of Multi-Agent Q-learning DynamicsAamal Abbas Hussain, Francesco Belardinelli, Georgios Piliouras. 1578-1586 [doi]
- Model-based Dynamic Shielding for Safe and Efficient Multi-agent Reinforcement LearningWenli Xiao, Yiwei Lyu, John M. Dolan. 1587-1596 [doi]
- Toward Risk-based Optimistic Exploration for Cooperative Multi-Agent Reinforcement LearningJihwan Oh, Joonkee Kim, Minchan Jeong, Se-Young Yun. 1597-1605 [doi]
- Counterexample-Guided Policy Refinement in Multi-Agent Reinforcement LearningBriti Gangopadhyay, Pallab Dasgupta, Soumyajit Dey. 1606-1614 [doi]
- Prioritized Tasks Mining for Multi-Task Cooperative Multi-Agent Reinforcement LearningYang Yu, Qiyue Yin, Junge Zhang, Kaiqi Huang. 1615-1623 [doi]
- M3: Modularization for Multi-task and Multi-agent Offline Pre-trainingLinghui Meng 0001, Jingqing Ruan, Xuantang Xiong, Xiyun Li, Xi Zhang, Dengpeng Xing, Bo Xu 0002. 1624-1633 [doi]
- User Device Interaction Prediction via Relational Gated Graph Attention Network and Intent-aware EncoderJingyu Xiao, Qingsong Zou, Qing Li 0004, Dan Zhao, Kang Li, Wenxin Tang, Runjie Zhou, Yong Jiang 0001. 1634-1642 [doi]
- Inferring Player Location in Sports Matches: Multi-Agent Spatial Imputation from Limited ObservationsGregory Everett, Ryan J. Beal, Tim Matthews, Joseph Early, Timothy J. Norman, Sarvapali D. Ramchurn. 1643-1651 [doi]
- Learning Graph-Enhanced Commander-Executor for Multi-Agent NavigationXinyi Yang, Shiyu Huang, Yiwen Sun, Yuxiang Yang, Chao Yu 0005, Wei-Wei Tu, Huazhong Yang, Yu Wang 0002. 1652-1660 [doi]
- Permutation-Invariant Set Autoencoders with Fixed-Size Embeddings for Multi-Agent LearningRyan Kortvelesy, Steven D. Morad, Amanda Prorok. 1661-1669 [doi]
- Infomaxformer: Maximum Entropy Transformer for Long Time-Series Forecasting ProblemPeiwang Tang, Xianchao Zhang. 1670-1678 [doi]
- TransfQMix: Transformers for Leveraging the Graph Structure of Multi-Agent Reinforcement Learning ProblemsMatteo Gallici, Mario Martin, Ivan Masmitja. 1679-1687 [doi]
- Intelligent Onboard Routing in Stochastic Dynamic Environments using TransformersRohit Chowdhury, Raswanth Murugan, Deepak N. Subramani. 1688-1696 [doi]
- Characterizations of Sequential Valuation RulesChris Dong, Patrick Lederer. 1697-1705 [doi]
- Collecting, Classifying, Analyzing, and Using Real-World Ranking DataNiclas Boehmer, Nathan Schaar. 1706-1715 [doi]
- Margin of Victory for Weighted Tournament SolutionsMichelle Döring, Jannik Peters 0001. 1716-1724 [doi]
- Bribery Can Get Harder in Structured Multiwinner Approval ElectionBartosz Kusek, Robert Bredereck, Piotr Faliszewski, Andrzej Kaczmarczyk 0001, Dusan Knop. 1725-1733 [doi]
- Strategyproof Social Decision Schemes on Super Condorcet DomainsFelix Brandt 0001, Patrick Lederer, Sascha Tausch. 1734-1742 [doi]
- Separating and Collapsing Electoral Control TypesBenjamin Carleton, Michael C. Chavrimootoo, Lane A. Hemaspaandra, David E. Narváez, Conor Taliancich, Henry B. Welles. 1743-1751 [doi]
- The Distortion of Approval Voting with RunoffSoroush Ebadian, Mohamad Latifian, Nisarg Shah 0001. 1752-1760 [doi]
- Models of Anxiety for Agent Deliberation: The Benefits of Anxiety-Sensitive AgentsArvid Horned, Loïs Vanhée. 1761-1767 [doi]
- Social Choice Around Decentralized Autonomous Organizations: On the Computational Social Choice of Digital CommunitiesNimrod Talmon. 1768-1773 [doi]
- Value Inference in Sociotechnical SystemsEnrico Liscio, Roger Lera-Leri, Filippo Bistaffa, Roel I. J. Dobbe, Catholijn M. Jonker, Maite López-Sánchez, Juan A. Rodríguez-Aguilar, Pradeep K. Murukannaiah. 1774-1780 [doi]
- Presenting Multiagent Challenges in Team Sports AnalyticsDavid Radke, Alexi Orchard. 1781-1785 [doi]
- Communication Meaning: Foundations and Directions for Systems ResearchAmit K. Chopra, Samuel H. Christie V.. 1786-1791 [doi]
- The Rule-Tool-User Nexus in Digital Collective DecisionsZoi Terzopoulou, Marijn A. Keijzer, Gogulapati Sreedurga, Jobst Heitzig. 1792-1796 [doi]
- Epistemic Side Effects: An AI Safety ProblemToryn Q. Klassen, Parand Alizadeh Alamdari, Sheila A. McIlraith. 1797-1801 [doi]
- Citizen-Centric Multiagent SystemsSebastian Stein 0001, Vahid Yazdanpanah. 1802-1807 [doi]
- FedMM: A Communication Efficient Solver for Federated Adversarial Domain AdaptationYan Shen, Jian Du, Han Zhao, Zhanghexuan Ji, Chunwei Ma, Mingchen Gao. 1808-1816 [doi]
- Adversarial Link Prediction in Spatial NetworksMichal Tomasz Godziszewski, Yevgeniy Vorobeychik, Tomasz P. Michalak. 1817-1825 [doi]
- Distributed Mechanism Design in Social NetworksHaoxin Liu, Yao Zhang, Dengji Zhao. 1826-1834 [doi]
- Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time AttacksMohammad Mohammadi, Jonathan Nöther, Debmalya Mandal, Adish Singla, Goran Radanovic. 1835-1844 [doi]
- How to Turn an MAS into a Graphical Causal ModelH. Van Dyke Parunak. 1845-1847 [doi]
- Differentiable Agent-based EpidemiologyAyush Chopra, Alexander Rodríguez, Jayakumar Subramanian, Arnau Quera-Bofarull, Balaji Krishnamurthy, B. Aditya Prakash, Ramesh Raskar. 1848-1857 [doi]
- Social Distancing via Social SchedulingDeepesh Kumar Lall, Garima Shakya, Swaprava Nath. 1858-1866 [doi]
- Don't Simulate Twice: One-Shot Sensitivity Analyses via Automatic DifferentiationArnau Quera-Bofarull, Ayush Chopra, Joseph Aylett-Bullock, Carolina Cuesta-Lázaro, Anisoara Calinescu, Ramesh Raskar, Michael J. Wooldridge. 1867-1876 [doi]
- Markov Aggregation for Speeding Up Agent-Based Movement SimulationsBernhard C. Geiger, Alireza Jahani, Hussain Hussain, Derek Groen. 1877-1885 [doi]
- Agent-Based Modeling of Human Decision-makers Under Uncertain Information During Supply Chain ShortagesNutchanon Yongsatianchot, Noah Chicoine, Jacqueline A. Griffin, Özlem Ergun, Stacy Marsella. 1886-1894 [doi]
- Simulating Panic Amplification in Crowds via Density-Emotion InteractionErik van Haeringen, Charlotte Gerritsen. 1895-1902 [doi]
- Modelling Agent Decision Making in Agent-based Simulation - Analysis Using an Economic Technology Uptake ModelFranziska Klügl, Hildegunn Kyvik Nordås. 1903-1911 [doi]
- Emotion Contagion in Agent-based Simulations of Crowds: A Systematic ReviewErik van Haeringen, Charlotte Gerritsen, Koen V. Hindriks. 1912-1914 [doi]
- Worst-Case Adaptive Submodular CoverJing Yuan 0002, ShaoJie Tang. 1915-1922 [doi]
- Minimax Strikes BackQuentin Cohen-Solal, Tristan Cazenave. 1923-1931 [doi]
- Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement LearningBram Grooten, Ghada Sokar, Shibhansh Dohare, Elena Mocanu, Matthew E. Taylor, Mykola Pechenizkiy, Decebal Constantin Mocanu. 1932-1941 [doi]
- Parameter Sharing with Network Pruning for Scalable Multi-Agent Deep Reinforcement LearningWoojun Kim, Youngchul Sung. 1942-1950 [doi]
- Learning Rewards to Optimize Global Performance Metrics in Deep Reinforcement LearningJunqi Qian, Paul Weng, Chenmien Tan. 1951-1960 [doi]
- A Deep Reinforcement Learning Approach for Online Parcel AssignmentHao Zeng, Qiong Wu 0005, Kunpeng Han, Junying He, Haoyuan Hu. 1961-1968 [doi]
- CoRaL: Continual Representation Learning for Overcoming Catastrophic ForgettingMohammad Samin Yasar, Tariq Iqbal. 1969-1978 [doi]
- Revealed Multi-Objective Utility Aggregation in Human DrivingAtrisha Sarkar, Kate Larson, Krzysztof Czarnecki 0001. 1979-1987 [doi]
- A Brief Guide to Multi-Objective Reinforcement Learning and PlanningConor F. Hayes, Roxana Radulescu, Eugenio Bargiacchi, Johan Källström, Matthew Macfarlane, Mathieu Reymond, Timothy Verstraeten, Luisa M. Zintgraf, Richard Dazeley, Fredrik Heintz, Enda Howley, Athirai A. Irissappane, Patrick Mannion, Ann Nowé, Gabriel de Oliveira Ramos, Marcello Restelli, Peter Vamplew 0001, Diederik M. Roijers. 1988-1990 [doi]
- Welfare and Fairness in Multi-objective Reinforcement LearningZimeng Fan 0002, Nianli Peng, Muhang Tian, Brandon Fain. 1991-1999 [doi]
- Preference-Based Multi-Objective Multi-Agent Path FindingFlorence Ho, Shinji Nakadai. 2000-2002 [doi]
- Sample-Efficient Multi-Objective Learning via Generalized Policy Improvement PrioritizationLucas Nunes Alegre, Ana L. C. Bazzan, Diederik M. Roijers, Ann Nowé, Bruno C. da Silva 0001. 2003-2012 [doi]
- MADDM: Multi-Advisor Dynamic Binary Decision-Making by Maximizing the UtilityZhaori Guo, Timothy J. Norman, Enrico H. Gerding. 2013-2021 [doi]
- On the Complexity of the Two-Stage Majority RuleYongjie Yang 0001. 2022-2030 [doi]
- Fairness in Participatory Budgeting via Equality of ResourcesJan Maly 0001, Simon Rey, Ulle Endriss, Martin Lackner. 2031-2039 [doi]
- Free-Riding in Multi-Issue DecisionsMartin Lackner, Jan Maly 0001, Oliviero Nardi. 2040-2048 [doi]
- k-Prize Weighted Voting GameWei-chen Lee, David Hyland, Alessandro Abate, Edith Elkind, Jiarui Gan, Julian Gutierrez 0001, Paul Harrenstein, Michael J. Wooldridge. 2049-2057 [doi]
- Computing the Best Policy that Survives a VoteAndrei Constantinescu, Roger Wattenhofer. 2058-2066 [doi]
- Voting by AxiomsMarie Christin Schmidtlein, Ulle Endriss. 2067-2075 [doi]
- A Hotelling-Downs Game for Strategic Candidacy with Binary IssuesJavier Maass, Vincent Mousseau, Anaëlle Wilczynski. 2076-2084 [doi]
- Voting with Limited Energy: A Study of Plurality and BordaZoi Terzopoulou. 2085-2093 [doi]
- Non-Obvious Manipulability for Single-Parameter Agents and Bilateral TradeThomas Archbold, Bart de Keijzer, Carmine Ventre. 2107-2115 [doi]
- Mechanism Design for Improving Accessibility to Public FacilitiesHau Chan, Chenhao Wang 0001. 2116-2124 [doi]
- Explicit Payments for Obviously Strategyproof MechanismsDiodato Ferraioli, Carmine Ventre. 2125-2133 [doi]
- Bilevel Entropy based Mechanism Design for Balancing Meta in Video GamesSumedh Pendurkar, Chris Chow, Luo Jie, Guni Sharon. 2134-2142 [doi]
- IQ-Flow: Mechanism Design for Inducing Cooperative Behavior to Self-Interested Agents in Sequential Social DilemmasBengisu Guresti, Abdullah Vanlioglu, Nazim Kemal Ure. 2143-2151 [doi]
- Settling the Distortion of Distributed Facility LocationAris Filos-Ratsikas, Panagiotis Kanellopoulos, Alexandros A. Voudouris, Rongsen Zhang. 2152-2160 [doi]
- Cost Sharing under Private Valuation and Connection ControlTianyi Zhang, Junyu Zhang, Sizhe Gu, Dengji Zhao. 2161-2169 [doi]
- Facility Location Games with ThresholdsHouyu Zhou, Guochuan Zhang, Lili Mei, Minming Li. 2170-2178 [doi]
- Random Majority Opinion Diffusion: Stabilization Time, Absorbing States, and Influential NodesAhad N. Zehmakan. 2179-2187 [doi]
- Axiomatic Analysis of Medial Centrality MeasuresWiktoria Kosny, Oskar Skibski. 2188-2196 [doi]
- Online Influence Maximization under Decreasing Cascade ModelFang Kong, Jize Xie, Baoxiang Wang 0001, Tao Yao, Shuai Li. 2197-2204 [doi]
- Node Conversion Optimization in Multi-hop Influence NetworksJie Zhang, Yuezhou Lv, Zihe Wang 0001. 2205-2212 [doi]
- Decentralized Core-periphery Structure in Social Networks Accelerates Cultural Innovation in Agent-based ModelingJesse Milzman, Cody Moser. 2213-2221 [doi]
- Being an Influencer is Hard: The Complexity of Influence Maximization in Temporal Graphs with a Fixed SourceArgyrios Deligkas, Eduard Eiben, Tiger-Lily Goldsmith, George Skretas. 2222-2230 [doi]
- Enabling Imitation-Based Cooperation in Dynamic Social NetworksJacques Bara, Paolo Turrini, Giulia Andrighetto. 2231-2233 [doi]
- The Grapevine Web: Analysing the Spread of False Information in Social Networks with Corrupted SourcesJacques Bara, Charlie Pilgrim, Paolo Turrini, Stanislav Zhydkov. 2234-2242 [doi]
- The Importance of Credo in Multiagent LearningDavid Radke, Kate Larson, Tim Brecht. 2243-2252 [doi]
- Contextual Integrity for Argumentation-based Privacy ReasoningGideon Ogunniye, Nadin Kökciyan. 2253-2261 [doi]
- Predicting Privacy Preferences for Smart Devices as NormsMarc Serramia, William Seymour, Natalia Criado, Michael Luck. 2262-2270 [doi]
- Agent-directed Runtime Norm SynthesisAndreasa Morris-Martin, Marina De Vos, Julian A. Padget, Oliver Ray. 2271-2279 [doi]
- Emergence of Norms in Interactions with Complex RewardsDhaminda B. Abeywickrama, Nathan Griffiths, Zhou Xu, Alex Mouzakitis. 2280-2282 [doi]
- Evaluating a Mechanism for Explaining BDI Agent BehaviourMichael Winikoff, Galina Sidorenko. 2283-2285 [doi]
- Learning Manner of Execution from Partial CorrectionsMattias Appelgren, Alex Lascarides. 2286-2288 [doi]
- What Do You Care About: Inferring Values from EmotionsJieting Luo, Mehdi Dastani, Thomas Studer, Beishui Liao. 2289-2291 [doi]
- 'Why didn't you allocate this task to them?' Negotiation-Aware Explicable Task Allocation and Contrastive Explanation GenerationZahra Zahedi, Sailik Sengupta, Subbarao Kambhampati. 2292-2294 [doi]
- Explaining Agent Preferences and Behavior: Integrating Reward Decomposition and Contrastive HighlightsYael Septon, Yotam Amitai, Ofra Amir. 2295-2297 [doi]
- Explanation Styles for Trustworthy Autonomous SystemsDavid A. Robb 0001, Xingkun Liu, Helen F. Hastie. 2298-2300 [doi]
- Modeling the Interpretation of Animations to Help Improve Emotional ExpressionTaíssa Ribeiro, Ricardo Rodrigues 0005, Carlos Martinho. 2301-2303 [doi]
- Artificial Prediction Markets Present a Novel Opportunity for Human-AI CollaborationTatiana Chakravorti, Vaibhav Singh, Sarah Rajtmajer, Michael McLaughlin, Robert Fraleigh, Christopher Griffin 0001, Anthony Kwasnica, David M. Pennock, C. Lee Giles. 2304-2306 [doi]
- Causal Explanations for Sequential Decision Making Under UncertaintySamer B. Nashed, Saaduddin Mahmud, Claudia V. Goldman, Shlomo Zilberstein. 2307-2309 [doi]
- Hierarchical Reinforcement Learning with Human-AI Collaborative Sub-Goals OptimizationHaozhe Ma, Thanh Vinh Vo, Tze-Yun Leong. 2310-2312 [doi]
- Context-aware Agents based on Psychological Archetypes for TeamworkAnupama Arukgoda, Erandi Lakshika, Michael Barlow, Kasun Gunawardana. 2313-2315 [doi]
- Personalized Agent Explanations for Human-Agent Teamwork: Adapting Explanations to User Trust, Workload, and PerformanceRuben S. Verhagen, Mark A. Neerincx, Can Parlar, Marin Vogel, Myrthe L. Tielman. 2316-2318 [doi]
- A Teachable Agent to Enhance Elderly's IkigaiPing Chen, Xinjia Yu, Sufang Lim, Zhiqi Shen 0001. 2319-2321 [doi]
- Improving Human-Robot Team Performance with Proactivity and Shared Mental ModelsGwendolyn Edgar, Matthew McWilliams, Matthias Scheutz. 2322-2324 [doi]
- Towards Explaining Sequences of Actions in Multi-Agent Deep Reinforcement Learning ModelsKhaing Phyo Wai, Minghong Geng, Budhitama Subagdja, Shubham Pateria, Ah-Hwee Tan. 2325-2327 [doi]
- Learning Constraints From Human Stop-Feedback in Reinforcement LearningSilvia Poletti, Alberto Testolin, Sebastian Tschiatschek. 2328-2330 [doi]
- Goal Alignment: Re-analyzing Value Alignment Problems Using Human-Aware AIMalek Mechergui, Sarath Sreedharan. 2331-2333 [doi]
- Effectiveness of Teamwork-Level Interventions through Decision-Theoretic Reasoning in a Minecraft Search-and-Rescue TaskDavid V. Pynadath, Nikolos Gurney, Sarah Kenny, Rajay Kumar, Stacy C. Marsella, Haley Matuszak, Hala Mostafa, Pedro Sequeira, Volkan Ustun, Peggy Wu. 2334-2336 [doi]
- Hierarchical Reinforcement Learning for Ad Hoc TeamingStéphane Aroca-Ouellette, Miguel Aroca-Ouellette, Upasana Biswas, Katharina Kann, Alessandro Roncone. 2337-2339 [doi]
- Asynchronous Communication Aware Multi-Agent Task AllocationBen Rachmut, Sofia Amador Nelke, Roie Zivan. 2340-2342 [doi]
- Towards Robust Contrastive Explanations for Human-Neural Multi-agent SystemsFrancesco Leofante, Alessio Lomuscio. 2343-2345 [doi]
- Visual Explanations for Defence in Abstract ArgumentationSylvie Doutre, Théo Duchatelle, Marie-Christine Lagasquie-Schiex. 2346-2348 [doi]
- Minimising Task Tardiness for Multi-Agent Pickup and DeliverySaravanan Ramanathan, Yihao Liu 0002, Xueyan Tang, Wentong Cai 0001, Jingning Li. 2349-2351 [doi]
- Probabilistic Deduction as a Probabilistic Extension of Assumption-based ArgumentationXiuyi Fan. 2352-2354 [doi]
- Bayes-Adaptive Monte-Carlo Planning for Type-Based Reasoning in Large Partially Observable, Multi-Agent EnvironmentsJonathon Schwartz, Hanna Kurniawati. 2355-2357 [doi]
- Blame Attribution for Multi-Agent Pathfinding Execution FailuresAvraham Natan, Roni Stern, Meir Kalech. 2358-2360 [doi]
- A Semantic Approach to Decidability in Epistemic PlanningAlessandro Burigana, Paolo Felli, Marco Montali, Nicolas Troquard. 2361-2363 [doi]
- Forward-PECVaR Algorithm: Exact Evaluation for CVaR SSPsWilly Arthur Silva Reis, Denis Benevolo Pais, Valdinei Freire, Karina Valdivia Delgado. 2364-2366 [doi]
- Explainable Ensemble Classification Model based on ArgumentationNadia Abchiche-Mimouni, Leila Amgoud, Farida Zehraoui. 2367-2369 [doi]
- Updating Action Descriptions and Plans for Cognitive AgentsPeter Stringer, Rafael C. Cardoso 0001, Clare Dixon, Michael Fisher 0001, Louise A. Dennis. 2370-2372 [doi]
- Argument-based Explanation FunctionsLeila Amgoud, Philippe Muller, Henri Trenquier. 2373-2375 [doi]
- A Formal Framework for Deceptive Topic Planning in Information-Seeking DialoguesAndreas Brännström, Virginia Dignum, Juan Carlos Nieves. 2376-2378 [doi]
- Memoryless Adversaries in Imperfect Information GamesDhananjay Raju, Georgios Bakirtzis, Ufuk Topcu. 2379-2381 [doi]
- Bounded and Unbounded Verification of RNN-Based Agents in Non-deterministic EnvironmentsMehran Hosseini, Alessio Lomuscio. 2382-2384 [doi]
- Methods and Mechanisms for Interactive Novelty Handling in Adversarial EnvironmentsTung Thai, Mudit Verma, Utkarsh Soni, Sriram Gopalakrishnan, Ming Shen, Mayank Garg, Ayush Kalani, Nakul Vaidya, Neeraj Varshney, Chitta Baral, Subbarao Kambhampati, Jivko Sinapov, Matthias Scheutz. 2385-2387 [doi]
- One-Shot Learning from a Demonstration with Hierarchical Latent LanguageNathaniel Weir, Xingdi Yuan, Marc-Alexandre Côté, Matthew J. Hausknecht, Romain Laroche, Ida Momennejad, Harm van Seijen, Benjamin Van Durme. 2388-2390 [doi]
- Emergent Compositional Concept Communication through Mutual Information in Multi-Agent TeamsSeth Karten, Siva Kailas, Katia P. Sycara. 2391-2393 [doi]
- Reasoning about Uncertainty in AgentSpeak using Dynamic Epistemic LogicMichael J. Vezina, François Schwarzentruber, Babak Esfandiari, Sandra Morley. 2394-2396 [doi]
- Towards Optimal and Scalable Evacuation Planning Using Data-driven Agent Based ModelsKazi Ashik Islam, Da Qi Chen, Madhav V. Marathe, Henning S. Mortveit, Samarth Swarup, Anil Vullikanti. 2397-2399 [doi]
- Intention Progression with Maintenance GoalsDi Wu, Yuan Yao, Natasha Alechina, Brian Logan 0001, John Thangarajah. 2400-2402 [doi]
- Safety Guarantees in Multi-agent Learning via Trapping RegionsAleksander Czechowski, Frans A. Oliehoek. 2403-2405 [doi]
- Multi-Team Fitness Critics For Robust TeamingJoshua Cook, Tristan Scheiner, Kagan Tumer. 2406-2408 [doi]
- Multi-Agent Deep Reinforcement Learning for High-Frequency Multi-Market MakingPankaj Kumar. 2409-2411 [doi]
- TA-Explore: Teacher-Assisted Exploration for Facilitating Fast Reinforcement LearningAli Beikmohammadi, Sindri Magnússon. 2412-2414 [doi]
- Which way is 'right'?: Uncovering limitations of Vision-and-Language Navigation ModelsMeera Hahn, Amit Raj, James M. Rehg. 2415-2417 [doi]
- Learning Individual Difference Rewards in Multi-Agent Reinforcement LearningChen Yang, Guangkai Yang, Junge Zhang. 2418-2420 [doi]
- TiLD: Third-person Imitation Learning by Estimating Domain Cognitive Differences of Visual DemonstrationsZixuan Chen, Wenbin Li, Yang Gao, Yiyu Chen. 2421-2423 [doi]
- Off-Beat Multi-Agent Reinforcement LearningWei Qiu 0001, Weixun Wang, Rundong Wang, Bo An 0001, Yujing Hu, Svetlana Obraztsova, Zinovi Rabinovich, Jianye Hao, Yingfeng Chen, Changjie Fan. 2424-2426 [doi]
- AJAR: An Argumentation-based Judging Agents Framework for Ethical Reinforcement LearningBenoît Alcaraz, Olivier Boissier, Rémy Chaput, Christopher Leturc. 2427-2429 [doi]
- Never Worse, Mostly Better: Stable Policy Improvement in Deep Reinforcement LearningPranav Khanna, Guy Tennenholtz, Nadav Merlis, Shie Mannor, Chen Tessler. 2430-2432 [doi]
- Selectively Sharing Experiences Improves Multi-Agent Reinforcement LearningMatthias Gerstgrasser, Tom Danino, Sarah Keren. 2433-2435 [doi]
- The Challenge of Redundancy on Multi-agent Value FactorisationSiddarth Shandeep Singh, Benjamin Rosman. 2436-2438 [doi]
- Robust Ordinal Regression for Collaborative Preference Learning with Opinion SynergiesHugo Gilbert, Mohamed Ouaguenouni, Meltem Öztürk, Olivier Spanjaard. 2439-2441 [doi]
- Off-the-Grid MARL: Datasets and Baselines for Offline Multi-Agent Reinforcement LearningClaude Formanek, Asad Jeewa, Jonathan P. Shock, Arnu Pretorius. 2442-2444 [doi]
- Search-Improved Game-Theoretic Multiagent Reinforcement Learning in General and Negotiation GamesZun Li, Marc Lanctot, Kevin R. McKee, Luke Marris, Ian Gemp, Daniel Hennes, Kate Larson, Yoram Bachrach, Michael P. Wellman, Paul Muller. 2445-2447 [doi]
- Grey-box Adversarial Attack on Communication in Multi-agent Reinforcement LearningXiao Ma, Wu-Jun Li. 2448-2450 [doi]
- Reward-Machine-Guided, Self-Paced Reinforcement LearningCevahir Köprülü, Ufuk Topcu. 2451-2453 [doi]
- Centralized Cooperative Exploration Policy for Continuous Control TasksChao Li, Chen Gong, Qiang He, Xinwen Hou, Yu Liu. 2454-2456 [doi]
- Do As You Teach: A Multi-Teacher Approach to Self-Play in Deep Reinforcement LearningChaitanya Kharyal, Tanmay Sinha, Sai Krishna Gottipati, Fatemeh Abdollahi, Srijita Das, Matthew E. Taylor. 2457-2459 [doi]
- PORTAL: Automatic Curricula Generation for Multiagent Reinforcement LearningJizhou Wu, Tianpei Yang, Xiaotian Hao, Jianye Hao, Yan Zheng, Weixun Wang, Matthew E. Taylor. 2460-2462 [doi]
- AI-driven Prices for Externalities and Sustainability in Production MarketsPanayiotis Danassis, Aris Filos-Ratsikas, Haipeng Chen 0001, Milind Tambe, Boi Faltings. 2463-2465 [doi]
- For One and All: Individual and Group Fairness in the Allocation of Indivisible GoodsJonathan Scarlett, Nicholas Teh, Yair Zick. 2466-2468 [doi]
- Matching Algorithms under Diversity-Based ReservationsHaris Aziz 0001, Sean Morota Chu, Zhaohong Sun 0001. 2469-2471 [doi]
- Social Mechanism Design: A Low-Level IntroductionBen Abramowitz, Nicholas Mattei. 2472-2474 [doi]
- Online 2-stage Stable MatchingEvripidis Bampis, Bruno Escoffier, Paul Youssef. 2475-2477 [doi]
- Strategic Play By Resource-Bounded Agents in Security GamesXinming Liu, Joseph Y. Halpern. 2478-2480 [doi]
- Neural Stochastic Agent-Based Limit Order Book Simulation: A Hybrid MethodologyZijian Shi, John Cartlidge. 2481-2483 [doi]
- Regularization for Strategy Exploration in Empirical Game-Theoretic AnalysisYongzhao Wang, Michael P. Wellman. 2484-2486 [doi]
- A Scalable Opponent Model Using Bayesian Learning for Automated Bilateral Multi-Issue NegotiationShengbo Chang, Katsuhide Fujita. 2487-2489 [doi]
- Benchmarking Robustness and Generalization in Multi-Agent Systems: A Case Study on Neural MMOYangkun Chen, Joseph Suarez, Junjie Zhang, ChengHui Yu, Bo Wu, HanMo Chen, Hengman Zhu, Rui Du, Shanliang Qian, Shuai Liu, Weijun Hong, Jinke He, Yibing Zhang, Liang Zhao, Clare Zhu, Julian Togelius, Sharada Mohanty, Jiaxin Chen, Xiu Li, Xiaolong Zhu, Phillip Isola. 2490-2492 [doi]
- SE4AI Issues on Social Media Agent Design with Use CasesFrancisco Supino Marcondes, José João Almeida, Paulo Novais. 2493-2495 [doi]
- Modeling Application Scenarios for Responsible Autonomy using Computational TranscendenceJayati Deshmukh, Nikitha Adivi, Srinath Srinivasa. 2496-2498 [doi]
- Domain-Expert Configuration of Hypermedia Multi-Agent Systems in Industrial Use CasesJérémy Lemée, Samuele Burattini, Simon Mayer, Andrei Ciortea. 2499-2501 [doi]
- Multi-Agent Reinforcement Learning for Fast-Timescale Demand Response of Residential LoadsVincent Mai, Philippe Maisonneuve, Tianyu Zhang, Hadi Nekoei, Liam Paull, Antoine Lesage-Landry. 2502-2504 [doi]
- The Swiss GambitÁgnes Cseh, Pascal Führlich, Pascal Lenzner. 2505-2507 [doi]
- An Adversarial Strategic Game for Machine Learning as a Service using System FeaturesGuoxin Sun, Tansu Alpcan, Seyit Camtepe, Andrew C. Cullen, Benjamin I. P. Rubinstein. 2508-2510 [doi]
- Optimizing Crop Management with Reinforcement Learning and Imitation LearningRan Tao, Pan Zhao, Jing Wu, Nicolas F. Martin, Matthew T. Harrison, Carla Sofia Santos Ferreira, Zahra Kalantari, Naira Hovakimyan. 2511-2513 [doi]
- A Novel Aggregation Framework for the Efficient Integration of Distributed Energy Resources in the Smart GridStavros Orfanoudakis, Georgios Chalkiadakis. 2514-2516 [doi]
- Near Optimal Strategies for Honeypots Placement in Dynamic and Large Active Directory NetworksHuy Q. Ngo, Mingyu Guo, Hung Nguyen 0004. 2517-2519 [doi]
- A Novel Demand Response Model and Method for Peak Reduction in Smart Grids - PowerTACSanjay Chandlekar, Arthik Boroju, Shweta Jain 0002, Sujit Gujar. 2520-2522 [doi]
- Robotic Shopping Assistance for Everyone: Dynamic Query Generation on a Semantic Digital Twin as a Basis for Autonomous Shopping AssistanceMichaela Kümpel, Jonas Dech, Alina Hawkin, Michael Beetz. 2523-2525 [doi]
- Counterfactually Fair Dynamic Assignment: A Case Study on PolicingTasfia Mashiat, Xavier Gitiaux, Huzefa Rangwala, Sanmay Das. 2526-2528 [doi]
- A Cloud-Based Solution for Multi-Agent Traffic Control SystemsChikadibia Ihejimba, Behnam Torabi, Rym Z. Wenkstern. 2529-2531 [doi]
- Balancing Fairness and Efficiency in Transport Network Design through Reinforcement LearningDimitris Michailidis, Sennay Ghebreab, Fernando P. Santos. 2532-2534 [doi]
- From Abstractions to Grounded Languages for Robust Coordination of Task Planning RobotsYu Zhang. 2535-2537 [doi]
- Idleness Estimation for Distributed Multiagent Patrolling StrategiesMehdi William Othmani-Guibourg, Jean-Loup Farges, Amal El Fallah-Seghrouchni. 2538-2540 [doi]
- Simpler rather than Challenging: Design of Non-Dyadic Human-Robot Collaboration to Mediate Human-Human Concurrent TasksFrancesco Semeraro, Jon Carberry, Angelo Cangelosi. 2541-2543 [doi]
- Learning to Self-Reconfigure for Freeform Modular Robots via Altruism Multi-Agent Reinforcement LearningLei Wu, Bin Guo 0001, Qiuyun Zhang, Zhuo Sun 0002, Jieyi Zhang, Zhiwen Yu 0001. 2544-2546 [doi]
- Learning Multiple Tasks with Non-stationary Interdependencies in Autonomous RobotsAlejandro Romero, Gianluca Baldassarre, Richard J. Duro, Vieri Giuliano Santucci. 2547-2549 [doi]
- Provably Manipulable 3D Structures using Graph TheoryJohn Harwell, London Lowmanstone, Maria L. Gini. 2550-2552 [doi]
- HoLA Robots: Mitigating Plan-Deviation Attacks in Multi-Robot Systems with Co-Observations and Horizon-Limiting AnnouncementsKacper Wardega, Max von Hippel, Roberto Tron, Cristina Nita-Rotaru, Wenchao Li 0001. 2553-2555 [doi]
- Online Re-Planning and Adaptive Parameter Update for Multi-Agent Path Finding with Stochastic Travel TimesAtsuyoshi Kita, Nobuhiro Suenari, Masashi Okada, Tadahiro Taniguchi. 2556-2558 [doi]
- RTransNav: Relation-wise Transformer Network for More Successful Object Goal NavigationKang Zhou, Chi Guo, Huyin Zhang, Wenfei Guo. 2559-2561 [doi]
- Multi-Agent Pickup and Delivery in Presence of Another Team of RobotsBenedetta Flammini, Davide Azzalini, Francesco Amigoni. 2562-2564 [doi]
- Reward Relabelling for Combined Reinforcement and Imitation Learning on Sparse-reward TasksJesus Bujalance Martin, Fabien Moutarde. 2565-2567 [doi]
- Connectivity Enhanced Safe Neural Network Planner for Lane Changing in Mixed TrafficXiangguo Liu, Ruochen Jiao, Bowen Zheng, Dave Liang, Qi Zhu 0002. 2568-2570 [doi]
- Bringing Diversity to Autonomous Vehicles: An Interpretable Multi-vehicle Decision-making and Planning FrameworkLicheng Wen, Pinlong Cai, Daocheng Fu, Song Mao, Yikang Li. 2571-2573 [doi]
- Loss of Distributed Coverage Using Lazy Agents Operating Under Discrete, Local, Event-Triggered CommunicationEdward Vickery, Aditya A. Paranjape. 2574-2576 [doi]
- Multi-Agent Path Finding via Reinforcement Learning with Hybrid RewardCheng Zhao, Liansheng Zhuang, Haonan Liu, Yihong Huang, Jian Yang. 2577-2579 [doi]
- Multi-Agent Pickup and Delivery with Task Probability DistributionAndrea Di Pietro, Nicola Basilico, Francesco Amigoni. 2580-2582 [doi]
- Minimally Constraining Line-of-Sight Connectivity Maintenance for Collision-free Multi-Robot Networks under UncertaintyYupeng Yang, Yiwei Lyu, Wenhao Luo. 2583-2585 [doi]
- Multi-Agent Path Finding with Time Windows: Preliminary ResultsJianqi Gao, Qi Liu, Shiyu Chen, Kejian Yan, Xinyi Li, Yanjie Li. 2586-2588 [doi]
- Two-Level Actor-Critic Using Multiple TeachersSu Zhang, Srijita Das, Sriram Ganapathi Subramanian, Matthew E. Taylor. 2589-2591 [doi]
- Provably Efficient Offline RL with OptionsXiaoyan Hu, Ho-Fung Leung. 2592-2594 [doi]
- Learning to Perceive in Deep Model-Free Reinforcement LearningGonçalo Querido, Alberto Sardinha, Francisco S. Melo. 2595-2597 [doi]
- SCRIMP: Scalable Communication for Reinforcement- and Imitation-Learning-Based Multi-Agent PathfindingYutong Wang, Bairan Xiang, Shinan Huang, Guillaume Sartoretti. 2598-2600 [doi]
- Learning Group-Level Information Integration in Multi-Agent CommunicationXiangrui Meng, Ying Tan. 2601-2603 [doi]
- Learnability with PAC Semantics for Multi-agent BeliefsIonela Georgiana Mocanu, Vaishak Belle, Brendan Juba. 2604-2606 [doi]
- Improving Cooperative Multi-Agent Exploration via Surprise Minimization and Social Influence MaximizationMingyang Sun, Yaqing Hou, Jie Kang, Haiyin Piao, Yifeng Zeng, Hongwei Ge, Qiang Zhang 0008. 2607-2609 [doi]
- Learning to Operate in Open Worlds by Adapting Planning ModelsWiktor Piotrowski, Roni Stern, Yoni Sher, Jacob Le, Matthew Klenk, Johan de Kleer, Shiwali Mohan. 2610-2612 [doi]
- End-to-End Optimization and Learning for Multiagent EnsemblesJames Kotary, Vincenzo Di Vito, Ferdinando Fioretto. 2613-2615 [doi]
- Optimal Decoy Resource Allocation for Proactive Defense in Probabilistic Attack GraphsHaoxiang Ma, Shuo Han 0002, Nandi Leslie, Charles A. Kamhoua, Jie Fu. 2616-2618 [doi]
- Referential Communication in Heterogeneous Communities of Pre-trained Visual Deep NetworksMateo Mahaut, Francesca Franzon, Roberto Dessì, Marco Baroni. 2619-2621 [doi]
- A Learning Approach to Complex Contagion Influence MaximizationHaipeng Chen 0001, Bryan Wilder, Wei Qiu 0001, Bo An 0001, Eric Rice, Milind Tambe. 2622-2624 [doi]
- Analyzing the Sensitivity to Policy-Value Decoupling in Deep Reinforcement Learning GeneralizationNasik Muhammad Nafi, Raja Farrukh Ali, William H. Hsu. 2625-2627 [doi]
- Reinforcement Learning with Depreciating AssetsTaylor Dohmen, Ashutosh Trivedi 0001. 2628-2630 [doi]
- Matching Options to Tasks using Option-Indexed Hierarchical Reinforcement LearningKushal Chauhan, Soumya Chatterjee 0002, Akash Reddy, Aniruddha S, Balaraman Ravindran, Pradeep Shenoy. 2631-2633 [doi]
- DGPO: Discovering Multiple Strategies with Diversity-Guided Policy OptimizationWenze Chen, Shiyu Huang, Yuan Chiang, Ting Chen, Jun Zhu. 2634-2636 [doi]
- Accelerating Neural MCTS Algorithms using Neural Sub-Net StructuresPrashank Kadam, Ruiyang Xu, Karl J. Lieberherr. 2637-2639 [doi]
- Provably Efficient Convergence of Primal-Dual Actor-Critic with Nonlinear Function ApproximationJing Dong 0008, Li Shen, Yinggan Xu, Baoxiang Wang 0001. 2640-2642 [doi]
- Achieving Near-optimal Regrets in Confounded Contextual BanditsXueping Gong, Jiheng Zhang. 2643-2645 [doi]
- Proportional Fairness in Obnoxious Facility LocationHaris Aziz 0001, Alexander Lam, Bo Li 0037, Fahimeh Ramezani, Toby Walsh. 2646-2648 [doi]
- Distortion in Attribute Approval Committee ElectionsDorothea Baumeister, Linus Boes. 2649-2651 [doi]
- Relaxations of Envy-Freeness Over GraphsJustin Payan, Rik Sengupta, Vignesh Viswanathan. 2652-2654 [doi]
- Fairly Allocating (Contiguous) Dynamic Indivisible Items with Few AdjustmentsMingWei Yang. 2655-2657 [doi]
- Measuring a Priori Voting Power - Taking Delegations SeriouslyRachael Colley, Théo Delemazure, Hugo Gilbert. 2658-2660 [doi]
- Sampling-Based Winner Prediction in District-Based ElectionsDebajyoti Kar, Palash Dey, Swagato Sanyal. 2661-2663 [doi]
- Cedric: A Collaborative DDoS Defense System Using CreditJiawei Li, Hui Wang, Jilong Wang 0001. 2664-2666 [doi]
- Social Aware Coalition Formation with Bounded Coalition SizeChaya Levinger, Amos Azaria, Noam Hazon. 2667-2669 [doi]
- Repeatedly Matching Items to Agents Fairly and EfficientlyIoannis Caragiannis, Shivika Narang. 2670-2672 [doi]
- The Complexity of Minimizing Envy in House AllocationJayakrishnan Madathil, Neeldhara Misra, Aditi Sethia. 2673-2675 [doi]
- Error in the Euclidean Preference ModelLuke Thorburn, Maria Polukarov, Carmine Ventre. 2676-2678 [doi]
- Distance Hypergraph Polymatrix Coordination GamesAlessandro Aloisio. 2679-2681 [doi]
- Search versus Search for Collapsing Electoral Control TypesBenjamin Carleton, Michael C. Chavrimootoo, Lane A. Hemaspaandra, David E. Narváez, Conor Taliancich, Henry B. Welles. 2682-2684 [doi]
- Does Delegating Votes Protect Against Pandering Candidates?Xiaolin Sun, Jacob Masur, Ben Abramowitz, Nicholas Mattei, Zizhan Zheng. 2685-2687 [doi]
- Resilient Fair Allocation of Indivisible GoodsDolev Mutzari, Yonatan Aumann, Sarit Kraus. 2688-2690 [doi]
- Stability of Weighted Majority Voting under Estimated WeightsShaojie Bai, Dongxia Wang 0002, Tim Muller, Peng Cheng, Jiming Chen 0001. 2691-2693 [doi]
- Indivisible Participatory Budgeting with Multiple Degrees of Sophistication of ProjectsGogulapati Sreedurga. 2694-2696 [doi]
- Incentivizing Sequential Crowdsourcing SystemsYuan Luo. 2697-2699 [doi]
- No-regret Learning Dynamics for Sequential Correlated EquilibriaHugh Zhang. 2700-2702 [doi]
- Fair Pricing for Time-Flexible Smart Energy MarketsRoland Saur, Han La Poutré, Neil Yorke-Smith. 2703-2705 [doi]
- Budget-Feasible Mechanism Design for Cost-Benefit Optimization in Gradual Service ProcurementFarzaneh Farhadi, Maria Chli, Nicholas R. Jennings. 2706-2708 [doi]
- Analysis of a Learning Based Algorithm for Budget PacingMohammadTaghi Hajiaghayi, Max Springer. 2709-2711 [doi]
- Finding Optimal Nash Equilibria in Multiplayer Games via Correlation PlansYouzhi Zhang 0001, Bo An 0001, V. S. Subrahmanian. 2712-2714 [doi]
- Diffusion Multi-unit Auctions with Diminishing Marginal Utility BuyersHaolin Liu, Xinyuan Lian, Dengji Zhao. 2715-2717 [doi]
- Improving Quantal Cognitive Hierarchy Model Through Iterative Population LearningYuhong Xu, Shih-Fen Cheng, Xinyu Chen. 2718-2720 [doi]
- A Nash-Bargaining-Based Mechanism for One-Sided Matching Markets and Dichotomous UtilitiesJugal Garg, Thorben Tröbst, Vijay V. Vazirani. 2721-2723 [doi]
- Differentially Private Diffusion Auction: The Single-unit CaseFengjuan Jia, Mengxiao Zhang, Jiamou Liu, Bakh Khoussainov. 2724-2726 [doi]
- Learning in Teams: Peer Evaluation for Fair Assessment of Individual ContributionsFedor Duzhin. 2727-2729 [doi]
- Agent-based Simulation of District-based Elections with Heterogeneous PopulationsAdway Mitra. 2730-2732 [doi]
- Deep Learning-based Spatially Explicit Emulation of an Agent-Based Simulator for Pandemic in a CityVarun Madhavan, Adway Mitra, Partha Pratim Chakrabarti. 2733-2735 [doi]
- A Decentralized Multiagent-Based Task Scheduling Framework for Handling Uncertain Events in Fog ComputingYikun Yang, Fenghui Ren, Minjie Zhang. 2736-2738 [doi]
- Co-evolution of Social and Non-social Guilt in Structured PopulationsTheodor Cimpeanu, Luís Moniz Pereira, The Anh Han. 2739-2741 [doi]
- Phantom - A RL-driven Multi-Agent Framework to Model Complex SystemsLeo Ardon, Jared Vann, Deepeka Garg, Thomas Spooner, Sumitra Ganesh. 2742-2744 [doi]
- Simulation Model with Side Trips at a Large-Scale EventRyo Niwa, Shunki Takami, Shusuke Shigenaka, Masaki Onishi, Wataru Naito, Tetsuo Yasutaka. 2745-2747 [doi]
- The Price of Algorithmic Pricing: Investigating Collusion in a Market Simulation with AI AgentsMichael Schlechtinger, Damaris Kosack, Heiko Paulheim, Thomas Fetzer, Franz Krause. 2748-2750 [doi]
- Crowd Simulation Incorporating a Route Choice Model and Similarity Evaluation using Real Large-scale DataRyo Nishida, Masaki Onishi, Koichi Hashimoto. 2751-2753 [doi]
- Capturing Hiders with Moving ObstaclesAyushman Panda, Kamalakar Karlapalem. 2754-2756 [doi]
- COBAI : A Generic Agent-based Model of Human Behaviors Centered on Contexts and InteractionsMaëlle Beuret, Irène Foucherot, Christian Gentil, Joël Savelli. 2757-2759 [doi]
- Learning Solutions in Large Economic Networks using Deep Multi-Agent Reinforcement LearningMichael Curry, Alexander Trott, Soham Phade, Yu Bai, Stephan Zheng. 2760-2762 [doi]
- Opinion Dynamics in Populations of Converging and Polarizing AgentsAnshul Toshniwal, Fernando P. Santos. 2763-2765 [doi]
- On a Voter Model with Context-Dependent Opinion AdoptionLuca Becchetti, Vincenzo Bonifaci, Emilio Cruciani, Francesco Pasquale. 2766-2768 [doi]
- Cognitive Bias-Aware Dissemination Strategies for Opinion Dynamics with External Information SourcesAbdullah Al Maruf, Luyao Niu, Bhaskar Ramasubramanian, Andrew Clark, Radha Poovendran. 2769-2771 [doi]
- Feature-based Individual Fairness in k-clusteringDebajyoti Kar, Mert Kosan, Debmalya Mandal, Sourav Medya, Arlei Silva, Palash Dey, Swagato Sanyal. 2772-2774 [doi]
- Fair Facility Location for Socially Equitable RepresentationHelen Sternbach, Sara Cohen. 2775-2777 [doi]
- S&F: Sources and Facts Reliability Evaluation MethodQuentin Elsaesser, Patricia Everaere, Sébastien Konieczny. 2778-2780 [doi]
- Offline Multi-Agent Reinforcement Learning with Coupled Value FactorizationXiangsen Wang, Xianyuan Zhan. 2781-2783 [doi]
- Learning Optimal "Pigovian Tax" in Sequential Social DilemmasYun Hua, Shang Gao, Wenhao Li, Bo Jin 0003, Xiangfeng Wang, Hongyuan Zha. 2784-2786 [doi]
- PACCART: Reinforcing Trust in Multiuser Privacy Agreement SystemsDaan Di Scala, Pinar Yolum. 2787-2789 [doi]
- Explain to Me: Towards Understanding Privacy DecisionsGonul Ayci, Arzucan Özgür, Murat Sensoy, Pinar Yolum. 2790-2791 [doi]
- The Resilience Game: A New Formalization of Resilience for Groups of Goal-Oriented Autonomous AgentsMichael A. Goodrich, Jennifer Leaf, Julie A. Adams, Matthias Scheutz. 2792-2794 [doi]
- Differentially Private Network Data Collection for Influence MaximizationM. Amin Rahimian, Fang-Yi Yu, Carlos Hurtado. 2795-2797 [doi]
- Inferring Implicit Trait Preferences from Demonstrations of Task Allocation in Heterogeneous TeamsVivek Mallampati, Harish Ravichandar. 2798-2800 [doi]
- From Scripts to RL Environments: Towards Imparting Commonsense Knowledge to RL AgentsAbhinav Joshi, Areeb Ahmad, Umang Pandey, Ashutosh Modi. 2801-2803 [doi]
- Hierarchical Reinforcement Learning with Attention RewardSihong Luo, Jinghao Chen, Zheng Hu, Chunhong Zhang, Benhui Zhuang. 2804-2806 [doi]
- Towards Multi-agent Learning of Causal NetworksStefano Mariani 0001, Pasquale Roseti, Franco Zambonelli. 2807-2809 [doi]
- FedHQL: Federated Heterogeneous Q-LearningFlint Xiaofeng Fan, Yining Ma 0001, Zhongxiang Dai, Cheston Tan, Bryan Kian Hsiang Low. 2810-2812 [doi]
- Know Your Enemy: Identifying and Adapting to Adversarial Attacks in Deep Reinforcement LearningSeán Caulfield Curley, Karl Mason, Patrick Mannion. 2813-2814 [doi]
- Transformer Actor-Critic with Regularization: Automated Stock Trading using Reinforcement LearningNamyeong Lee, Jun Moon. 2815-2817 [doi]
- Model-Based Actor-Critic for Multi-Objective Reinforcement Learning with Dynamic Utility FunctionsJohan Källström, Fredrik Heintz. 2818-2820 [doi]
- Relaxed Exploration Constrained Reinforcement LearningShahaf S. Shperberg, Bo Liu, Peter Stone. 2821-2823 [doi]
- Causality Detection for Efficient Multi-Agent Reinforcement LearningRafael Pina, Varuna De Silva, Corentin Artaud. 2824-2826 [doi]
- Diversity Through Exclusion (DTE): Niche Identification for Reinforcement Learning through Value-DecompositionPeter Sunehag, Alexander Sasha Vezhnevets, Edgar A. Duéñez-Guzmán, Igor Mordatch, Joel Z. Leibo. 2827-2829 [doi]
- Temporally Layered Architecture for Adaptive, Distributed and Continuous ControlDevdhar Patel, Joshua Russell, Francesca Walsh, Tauhidur Rahman, Terrence J. Sejnowski, Hava T. Siegelmann. 2830-2832 [doi]
- Multi-objective Reinforcement Learning in Factored MDPs with Graph Neural NetworksMarc Vincent, Amal El Fallah-Seghrouchni, Vincent Corruble, Narayan Bernardin, Rami Kassab, Frédéric Barbaresco. 2833-2835 [doi]
- An Analysis of Connections Between Regret Minimization and Actor Critic Methods in Cooperative SettingsChirag Chhablani, Ian A. Kash. 2836-2838 [doi]
- Attention-Based Recurrency for Multi-Agent Reinforcement Learning under State UncertaintyThomy Phan, Fabian Ritz, Jonas Nüßlein, Michael Kölle 0001, Thomas Gabor, Claudia Linnhoff-Popien. 2839-2841 [doi]
- A Theory of Mind Approach as Test-Time Mitigation Against Emergent Adversarial CommunicationNancirose Piazza, Vahid Behzadan. 2842-2844 [doi]
- Defensive Collaborative Learning: Protecting Objective Privacy in Data SharingCynthia Huang, Pascal Poupart. 2845-2847 [doi]
- Neuro-Symbolic World Models for Adapting to Open World NoveltyJonathan C. Balloch, Zhiyu Lin, Xiangyu Peng, Mustafa Hussain, Aarun Srinivas, Robert Wright, Julia M. Kim, Mark O. Riedl. 2848-2850 [doi]
- Modeling Dynamic Environments with Scene Graph MemoryAndrey Kurenkov, Michael Lingelbach, Tanmay Agarwal, Chengshu Li 0002, Emily Jin, Ruohan Zhang, Li Fei-Fei 0001, Jiajun Wu 0001, Silvio Savarese, Roberto Martín-Martín. 2851-2853 [doi]
- Group Fair Clustering Revisited - Notions and Efficient AlgorithmShivam Gupta 0004, Ganesh Ghalme, Narayanan C. Krishnan, Shweta Jain 0002. 2854-2856 [doi]
- LTL-Based Non-Markovian Inverse Reinforcement LearningMohammad Afzal, Sankalp Gambhir, Ashutosh Gupta, Krishna S, Ashutosh Trivedi 0001, Alvaro Velasquez. 2857-2859 [doi]
- The Parameterized Complexity of Welfare Guarantees in Schelling SegregationArgyrios Deligkas, Eduard Eiben, Tiger-Lily Goldsmith. 2860-2862 [doi]
- Fair Chore Division under Binary Supermodular CostsSiddharth Barman, Vishnu V. Narayan, Paritosh Verma. 2863-2865 [doi]
- Deliberation as Evidence Disclosure: A Tale of Two Protocol TypesJulian Chingoma, Adrian Haret. 2866-2868 [doi]
- How Does Fairness Affect the Complexity of Gerrymandering?Sandip Banerjee, Rajesh Chitnis, Abhiruk Lahiri. 2869-2871 [doi]
- Individual-Fair and Group-Fair Social Choice Rules under Single-Peaked PreferencesGogulapati Sreedurga, Soumyarup Sadhukhan, Souvik Roy, Yadati Narahari. 2872-2874 [doi]
- Maximin Share Allocations for Assignment ValuationsPooja Kulkarni, Rucha Kulkarni, Ruta Mehta. 2875-2876 [doi]
- Computational Complexity of Verifying the Group No-show ParadoxFarhad Mohsin, Qishen Han, Sikai Ruan, Pin-Yu Chen, Francesca Rossi, Lirong Xia. 2877-2879 [doi]
- Optimal Capacity Modification for Many-To-One Matching ProblemsJiehua Chen 0001, Gergely Csáji. 2880-2882 [doi]
- Learning to Explain Voting RulesInwon Kang, Qishen Han, Lirong Xia. 2883-2885 [doi]
- MMS Allocations of Chores with Connectivity Constraints: New Methods and New ResultsMingyu Xiao, Guoliang Qiu 0001, Sen Huang. 2886-2888 [doi]
- Group Fairness in Peer ReviewHaris Aziz 0001, Evi Micha, Nisarg Shah 0001. 2889-2891 [doi]
- Altruism in Facility Location ProblemsHouyu Zhou, Hau Chan, Minming Li. 2892-2894 [doi]
- Transfer Learning based Agent for Automated NegotiationSiqi Chen, Qisong Sun, Heng You, Tianpei Yang, Jianye Hao. 2895-2898 [doi]
- Single-Peaked Jump Schelling GamesTobias Friedrich 0001, Pascal Lenzner, Louise Molitor, Lars Seifert. 2899-2901 [doi]
- Defining Deception in Structural Causal GamesFrancis Rhys Ward, Francesca Toni, Francesco Belardinelli. 2902-2904 [doi]
- Game Model Learning for Mean Field GamesYongzhao Wang, Michael P. Wellman. 2905-2907 [doi]
- Modeling Robustness in Decision-Focused Learning as a Stackelberg GameSonja Johnson-Yu, Kai Wang 0040, Jessie Finocchiaro, Aparna Taneja, Milind Tambe. 2908-2909 [doi]
- Two-phase Security GamesAndrzej Nagórko, Pawel Ciosmak, Tomasz P. Michalak. 2910-2912 [doi]
- Stationary Equilibrium of Mean Field Games with Congestion-dependent Sojourn TimesCostas Courcoubetis, Antonis Dimakis. 2913-2915 [doi]
- Last-mile Collaboration: A Decentralized Mechanism with Performance Guarantees and its ImplementationKeyang Zhang, Jose Javier Escribano Macias, Dario Paccagnan, Panagiotis Angeloudis. 2916-2918 [doi]
- Deep Learning-Powered Iterative Combinatorial Auctions with Active LearningBenjamin Estermann, Stefan Kramer, Roger Wattenhofer, Ye Wang. 2919-2921 [doi]
- Revenue Maximization Mechanisms for an Uninformed Mediator with Communication AbilitiesZhikang Fan, Weiran Shen. 2922-2924 [doi]
- Counterfactual Explanations for Reinforcement Learning AgentsJasmina Gajcin. 2925-2927 [doi]
- Bipartite Matching for Repeated Allocation ProblemsYohai Trabelsi. 2928-2930 [doi]
- Artificial Intelligence Algorithms for Strategic Reasoning over Complex Multiagent SystemsZun Li. 2931-2933 [doi]
- Emergence of Cooperation on NetworksJacques Bara. 2934-2936 [doi]
- Enhancing User Understanding of Reinforcement Learning Agents Through Visual ExplanationsYotam Amitai. 2937-2939 [doi]
- Algorithmic Fairness in Temporal Resource AllocationAshwin Kumar. 2940-2942 [doi]
- AI & Multi-agent Systems for Data-centric Epidemic ForecastingAlexander Rodríguez. 2943-2945 [doi]
- Strategy Extraction for Transfer in AI AgentsArchana Vadakattu. 2946-2948 [doi]
- Multi-Advisor Dynamic Decision MakingZhaori Guo. 2949-2951 [doi]
- Forward-Looking and Backward-Looking Responsibility Attribution in Multi-Agent Sequential Decision MakingStelios Triantafyllou. 2952-2954 [doi]
- Coalition Formation in Sequential Decision-Making under UncertaintySaar Cohen 0001. 2955-2957 [doi]
- Fine Grained Complexity of Fair and Efficient AllocationsAditi Sethia. 2958-2960 [doi]
- Preference Inference from Demonstration in Multi-objective Multi-agent Decision MakingJunlin Lu. 2961-2963 [doi]
- Explanation through Dialogue for Reasoning SystemsYifan Xu. 2964-2966 [doi]
- Logics for Information AggregationJohn Lindqvist. 2967-2969 [doi]
- Towards Sample-Efficient Multi-Objective Reinforcement LearningLucas Nunes Alegre. 2970-2972 [doi]
- Verifiably Safe Decision-Making for Autonomous SystemsYi Yang. 2973-2975 [doi]
- A Toolkit for Encouraging Safe Diversity in Skill DiscoveryMaxence Hussonnois. 2976-2978 [doi]
- Citizen Centric Demand Responsive TransportAlexander Masterman. 2979-2981 [doi]
- Safe Behavior Specification and Planning for Autonomous Robotic Systems in Uncertain EnvironmentsJan Vermaelen. 2982-2984 [doi]
- Mechanism Design for Heterogeneous and Distributed Facility Location ProblemsRongsen Zhang. 2985-2987 [doi]
- Reinforcement Learning and Mechanism Design for Routing of Connected and Autonomous VehiclesBehrad Koohy. 2988-2990 [doi]
- Uncertainty-aware Personal Assistant and Explanation Method for Privacy DecisionsGönül Ayci. 2991-2992 [doi]
- Fair Transport Network Design using Multi-Agent Reinforcement LearningDimitris Michailidis. 2993-2995 [doi]
- Towards Scalable and Robust Decision Making in Partially Observable, Multi-Agent EnvironmentsJonathon Schwartz. 2996-2998 [doi]
- Reinforcement Learning in Multi-Objective Multi-Agent SystemsWillem Röpke. 2999-3001 [doi]
- Characterizing Fairness in Societal Resource AllocationTasfia Mashiat. 3002-3004 [doi]
- Learning Transferable Representations for Non-stationary EnvironmentsMohammad Samin Yasar. 3005-3007 [doi]
- Effective Human-Machine Teaming through Communicative Autonomous Agents that Explain, Coach, and ConvinceAaquib Tabrez. 3008-3010 [doi]
- Towards a Logical Account for Human-Aware Explanation Generation in Model Reconciliation ProblemsStylianos Loukas Vasileiou. 3011-3013 [doi]
- Contests and Other Topics in Multi-Agent SystemsAbheek Ghosh. 3014-3016 [doi]
- Planning and Coordination for Unmanned Aerial VehiclesJonathan Diller. 3017-3019 [doi]
- Towards Creating Better Interactive Agents: Leveraging Both Implicit and Explicit Human FeedbackKate Candon. 3020-3022 [doi]
- Assistive Robotics for Empowering Humans with Visual Impairments to Independently Perform Day-to-day TasksShivendra Agrawal. 3023-3025 [doi]
- Separations and Collapses in Computational Social ChoiceMichael C. Chavrimootoo. 3026-3028 [doi]
- Emergent Responsible Autonomy in Multi-Agent SystemsJayati Deshmukh. 3029-3031 [doi]
- Learning Representations and Robust Exploration for Improved Generalization in Reinforcement LearningNasik Muhammad Nafi. 3032-3034 [doi]
- Enhancing Smart, Sustainable Mobility with Game Theory and Multi-Agent Reinforcement LearningLucia Cipolina-Kun. 3035-3037 [doi]
- TDD for AOP: Test-Driven Development for Agent-Oriented ProgrammingCleber Jorge Amaral, Jomi Fred Hübner, Timotheus Kampik. 3038-3040 [doi]
- Interaction-Oriented Programming: Intelligent, Meaning-Based Multiagent SystemsAmit K. Chopra, Samuel H. Christie V., Munindar P. Singh. 3041-3043 [doi]
- Improvement and Evaluation of the Policy Legibility in Reinforcement LearningYanyu Liu, Yifeng Zeng, Biyang Ma, Yinghui Pan, Huifan Gao, Xiaohan Huang. 3044-3046 [doi]
- Multi-Robot Warehouse Optimization: Leveraging Machine Learning for Improved PerformanceMara Cairo, Bevin Eldaphonse, Payam Mousavi, Sahir, Sheikh Jubair, Matthew E. Taylor, Graham Doerksen, Nikolai Kummer, Jordan Maretzki, Gupreet Mohhar, Sean Murphy, Johannes Gunther, Laura Petrich, Talat Syed. 3047-3049 [doi]
- Robust JaCaMo Applications via Exceptions and AccountabilityMatteo Baldoni, Cristina Baroglio, Roberto Micalizio, Stefano Tedeschi 0001. 3050-3052 [doi]
- A Web-based Tool for Detecting Argument Validity and NoveltySandrine Chausson, Ameer Saadat-Yazdi, Xue Li, Jeff Z. Pan, Vaishak Belle, Nadin Kökciyan, Björn Ross. 3053-3055 [doi]
- Visualizing Logic Explanations for Social Media ModerationMarc Roig Vilamala, Dave Braines, Federico Cerutti 0001, Alun D. Preece. 3056-3058 [doi]
- The Influence Maximisation GameSukankana Chakraborty, Sebastian Stein 0001, Ananthram Swami, Matthew Jones, Lewis Hill. 3059-3061 [doi]
- Demonstrating Performance Benefits of Human-Swarm TeamingWilliam Hunt, Jack Ryan, Ayodeji Opeyemi Abioye, Sarvapali D. Ramchurn, Mohammad Divband Soorati. 3062-3064 [doi]
- Hiking up that HILL with Cogment-Verse: Train & Operate Multi-agent Systems Learning from HumansSai Krishna Gottipati, Luong Ha Nguyen, Clodéric Mars, Matthew E. Taylor. 3065-3067 [doi]
- Real Time Gesturing in Embodied Agents for Dynamic Content CreationHazel Watson-Smith, Felix Marcon Swadel, Jo Hutton, Kirstin Marcon, Mark Sagar, Shane Blackett, Tiago Rebeiro, Travers Biddle, Tim Wu 0002. 3068-3069 [doi]