Instructions
Below is the schedule of poster presentations by the authors of accepted extended abstracts. The recommended poster size is A0, vertical. The poster will be displayed all day on the day given in the schedule, and presented during the lunch and coffee breaks.
Authors of full papers will also have an opportunity to present a poster as well. Poster boards 49A-88B will be available on the day of the full paper presentation. We will publish allocations of boards to full papers later in April.
Print Posters On-site
Authors can print their posters on-site through a third-party service from Warehouse Stationery NZ at their Newmarket Branch (this is the nearest one to the conference location; click here to see the location on Google Maps). You can print posters either by walking in (offline) or making an order online by following the steps below.
- Go to the website of Warehouse Stationery NZ: https://www.wsprint.co.nz/value-poster/
- Set Size / Paper Type: A0 150 gsm
- Set Quantity: 1
- Set Scale to fit: Scale to fit
- Set: No Lamination
- Upload your poster PDF file
- Proceed by clicking “Check out” and then “Save and add to cart”
- Choose “Shipping” or “Click and Collect” Location: Click and Collect at the Warehouse Stationery Newmarket Branch
Map of the Great Room with Poster Boards
Poster Schedule for Full Papers
Each session is assigned to a range of spatially close-together boards. You may put your poster on any board within the range.
Wednesday 8 May
Sessions | Boards |
RL1, RL2, RL3, LEARN1 | 49A – 61A |
KR1, KR2 | 67A – 73A |
SC1, SC2 | 73B – 75B, 83A-87A |
COIN | 78A – 80B |
APP1 | 81A-82B, 87B-88B |
HUMAN1 | 61B – 64B |
EMAS | 65A-66B, 76A-77B |
Thursday 9 May
Sessions | Boards |
RL4, RL5, RL6, RL7 | 49A-61A |
KR3, KR4 | 67A – 73A |
SC3, SC4, MA1, MA2 | 73B – 75B, 77B – 82B, 83A-87B, 88A, 88B |
HUMAN2 | 61B – 64B |
SIM1 | 65A-66B, 76A-77B |
Friday 10 May
Sessions | Boards |
RL8, RL9, RL10, LEARN2 | 49A-62B |
SC5, SC6, MA3 | 73B – 75B, 80A-82B, 83A-87B |
SIM2, SIM3 | 67A – 73A |
ROBOT | 63B – 66B |
APP+ROB+PLAN | 76A – 78B |
Poster Schedule for Extended Abstracts
ID | Title | Author(s) | Day | Board |
3 | Pure Nash Equilibria in Weighted Congestion Games with Complementarities and Beyond | Kenjiro Takazawa | 08-May | 21A |
8 | Continual Depth-limited Responses for Computing Counter-strategies in Sequential Games | David Milec, Ondřej Kubíček and Viliam Lisy | 08-May | 10B |
12 | MATLight: Traffic Signal Coordinated Control Algorithm Based on Heterogeneous-Agent Mirror Learning With Transformer | Haipeng Zhang, Zhiwen Wang and Na Li | 08-May | 13a |
13 | Liquid Democracy for Low-Cost Ensemble Pruning | Ben Armstrong and Kate Larson | 08-May | 22A |
14 | Gaze Supervision for Mitigating Causal Confusion in Driving Agents | Abhijat Biswas, Badal Arun Pardhi, Caleb Chuck, Jarrett Holtz, Scott Niekum, Henny Admoni and Alessandro Allievi | 08-May | 1A |
21 | Emergent Dominance Hierarchies in Reinforcement Learning Agents | Ram Rachum, Yonatan Nakar, Bill Tomlinson, Nitay Alon and Reuth Mirsky | 08-May | 14A |
32 | Multi-level aggregation with delays and stochastic arrivals | Mathieu Mari, Michał Pawłowski, Runtian Ren and Piotr Sankowski | 08-May | 27A |
41 | Leveraging Approximate Model-based Shielding for Probabilistic Safety Guarantees in Continuous Environments | Alexander W. Goodall and Francesco Belardinelli | 08-May | 43B |
50 | Strategic Routing and Scheduling for Evacuations | Kazi Ashik Islam, Da Qi Chen, Madhav Marathe, Henning Mortveit, Samarth Swarup and Anil Vullikanti | 08-May | 11B |
67 | Shield Decentralization for Safe Reinforcement Learning in General Partially Observable Multi-Agent Environments | Daniel Melcer, Christopher Amato and Stavros Tripakis | 08-May | 28A |
77 | Clique Analysis and Bypassing in Continuous-Time Conflict-Based Search | Thayne T. Walker, Nathan Sturtevant and Ariel Felner | 08-May | 29A |
85 | Towards Understanding How to Reduce Generalization Gap in Visual Reinforcement Learning | Jiafei Lyu, Le Wan, Xiu Li and Zongqing Lu | 08-May | 37A |
94 | Attila: a Negotiating Agent for the Game of Diplomacy, Based on Purely Symbolic A.I. | Dave de Jonge and Laura Rodriguez Cima | 08-May | 12B |
107 | From Explicit Communication to Tacit Cooperation: A Novel Paradigm for Cooperative MARL | Dapeng Li, Zhiwei Xu, Bin Zhang, Guangchong Zhou, Zeren Zhang and Guoliang Fan | 08-May | 38A |
108 | Evaluation of Robustness of Off-Road Autonomous Driving Segmentation against Adversarial Attacks: A Dataset-Centric Study | Pankaj Deoli, Rohit Kumar, Axel Vierling and Karsten Berns | 09-May | 4B |
113 | Simulated Robotic Soft Body Manipulation | Glareh Mir and Michael Beetz | 08-May | 28B |
114 | MA-MIX: Value Function Decomposition for Cooperative Multiagent Reinforcement Learning Based on Multi-Head Attention Mechanism | Yu Niu, Hengxu Zhao and Lei Yu | 08-May | 39A |
115 | A Comparison of the Myerson Value and the Position Value | Ayşe Mutlu Derya | 08-May | 23A |
117 | A Negotiator’s Backup Plan: Optimal Concessions with a Reservation Value | Tamara C.P. Florijn, Pinar Yolum and Tim Baarslag | 08-May | 13B |
140 | Decentralised Emergence of Robust and Adaptive Linguistic Conventions in Populations of Autonomous Agents Grounded in Continuous Worlds | Jérôme Botoko Ekila, Jens Nevens, Lara Verheyen, Katrien Beuls and Paul Van Eecke | 08-May | 27B |
142 | Benchmarking MARL on Long Horizon Sequential Multi-Objective Tasks | Minghong Geng, Shubham Pateria, Budhitama Subagdja and Ah-Hwee Tan | 09-May | 1B |
143 | Fair Scheduling of Indivisible Chores | Yatharth Kumar, Sarfaraz Equbal, Rohit Gurjar, Swaprava Nath and Rohit Vaish | 08-May | 24A |
152 | TIMAT: Temporal Information Multi-Agent Transformer | Qitong Kang, Fuyong Wang, Zhongxin Liu and Zengqiang Chen | 08-May | 40B |
154 | Momentum-based Algorithm on Deep Reinforcement Learning | Huihui Zhang | 08-May | 39B |
158 | Contiguous Allocation of Binary Valued Indivisible Items on a Path | Yasushi Kawase, Bodhayan Roy and Mohammad Azharuddin Sanpui | 08-May | 14B |
166 | Decentralized Control of Distributed Manipulators: An Information Diffusion Approach | Nicolas Bessone, Payam Zahadat and Kasper Stoy | 09-May | 3B |
175 | Charging Electric Vehicles Fairly and Efficiently | Ramsundar Anandanarayanan, Swaprava Nath and Rohit Vaish | 08-May | 24B |
180 | Auto-Encoding Adversarial Imitation Learning | Kaifeng Zhang, Rui Zhao, Ziming Zhang and Yang Gao | 08-May | 38B |
190 | Ethical Markov Decision Processes with Moral Worth as Rewards | Mihail Stojanovski, Nadjet Bourdache, Grégory Bonnet and Mouaddib Abdel-Illah | 08-May | 30A |
197 | Game Transformations That Preserve Nash Equilibria or Best Response Sets | Emanuel Tewolde and Vincent Conitzer | 08-May | 15B |
213 | Mastering Robot Control through Point-based Reinforcement Learning with Pre-training | Yihong Chen, Cong Wang, Tianpei Yang, Meng Wang, Yingfeng Chen, Jifei Zhou, Chaoyi Zhao, Xinfeng Zhang, Zeng Zhao, Changjie Fan, Zhipeng Hu, Rong Xiong and Long Zeng | 08-May | 37B |
215 | On the Utility of External Agent Intention Predictor for Human-AI Coordination | Chenxu Wang, Zilong Chen and Huaping Liu | 08-May | 2A |
219 | A SAT-based Approach for Argumentation Dynamics | Jean Marie Lagniez, Emmanuel Lonca and Jean-Guy Mailly | 08-May | 31A |
221 | Optimal Diffusion Auctions | Yao Zhang, Shanshan Zheng and Dengji Zhao | 08-May | 17A |
222 | Attention Graph for Multi-Robot Social Navigation with Deep Reinforcement Learning | Erwan Escudie, Laetitia Matignon and Jacques Saraydaryan | 09-May | 2B |
225 | DCT: Dual Channel Training of Action Embeddings for Reinforcement Learning with Large Discrete Action Spaces | Pranavi Pathakota, Hardik Meisheri and Harshad Khadilkar | 08-May | 36B |
226 | Decision Market Based Learning For Multi-agent Contextual Bandit Problems | Wenlong Wang and Thomas Pfeiffer | 08-May | 35B |
235 | ENOTO: Improving Offline-to-Online Reinforcement Learning with Q-Ensembles | Kai Zhao, Jianye Hao, Yi Ma, Jinyi Liu, Yan Zheng and Zhaopeng Meng | 08-May | 34B |
238 | Computational Theory of Mind with Abstractions for Effective Human-Agent Collaboration | Emre Erdogan, Rineke Verbrugge and Pinar Yolum | 08-May | 3A |
239 | Centralized Training with Hybrid Execution in Multi-Agent Reinforcement Learning | Pedro P. Santos, Diogo Carvalho, Miguel Vasco, Alberto Sardinha, Pedro A. Santos, Ana Paiva and Francisco Melo | 08-May | 41A |
241 | Computing Balanced Solutions for Large International Kidney Exchange Schemes When Cycle Length Is Unbounded | Márton Benedek, Péter Biró, Gergely Csáji, Matthew Johnson, Daniël Paulusma and Xin Ye | 08-May | 23B |
249 | Projection-Optimal Monotonic Value Function Factorization in Multi-Agent Reinforcement Learning | Yongsheng Mei, Hanhan Zhou and Tian Lan | 08-May | 42A |
255 | Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent Deep Reinforcement Learning | Maxime Toquebiau, Nicolas Bredeche, Faïz Ben Amar and Jae-Yun Jun | 08-May | 43A |
256 | Decentralized Safe Control for Multi-Robot Navigation in Dynamic Environments with Limited Sensing | Saad Khan, Mayank Baranwal and Srikant Sukumar | 09-May | 10A |
257 | Synthesizing social laws with ATL conditions | Rustam Galimullin and Louwe B. Kuijer | 08-May | 32A |
261 | Deep Learning for Two-Sided Matching Markets | Sai Srivatsa Ravindranath, Zhe Feng, Shira Li, Jonathan Ma, Scott Kominers and David Parkes | 08-May | 22B |
262 | Combinatorial Client-Master Multiagent Deep Reinforcement Learning for Task Offloading in Mobile Edge Computing | Tesfay Zemuy Gebrekidan, Sebastian Stein and Timothy Norman | 08-May | 44A |
268 | Enhancing Search and Rescue Capabilities in Hazardous Communication-Denied Environments through Path-Based Sensors with Backtracking | Alexander Mendelsohn, Donald Sofge and Michael Otte | 08-May | 32B |
272 | Entropy Seeking Constrained Multiagent Reinforcement Learning | Ayhan Alp Aydeniz, Enrico Marchesini, Christopher Amato and Kagan Tumer | 08-May | 45A |
283 | Indirect Credit Assignment in a Multiagent System” | Everardo Gonzalez, Siddarth Viswanathan and Kagan Tumer | 08-May | 26B |
290 | Influence-Focused Asymmetric Island Model | Andrew Festa, Gaurav Dixit and Kagan Tumer | 08-May | 33Q |
298 | PADDLE: Logic Program Guided Policy Reuse in Deep Reinforcement Learning | Hao Zhang, Tianpei Yang, Yan Zheng, Jianye Hao and Matthew E. Taylor | 08-May | 46A |
301 | NP^PP-Completeness of Control by Adding Players to Change the Penrose–Banzhaf Power Index in Weighted Voting Games | Joanna Kaczmarek and Jörg Rothe | 08-May | 21B |
305 | Fairness of Exposure in Online Restless Multi-armed Bandits | Archit Sood, Shweta Jain and Sujit Gujar | 08-May | 47A |
307 | Analyzing Crowdfunding of Public Projects Under Dynamic Beliefs | Sankarshan Damle and Sujit Gujar | 08-May | 18A |
312 | No Transaction Fees? No Problem! Achieving Fairness in Transaction Fee Mechanism Design | Sankarshan Damle, Varul Srivastava and Sujit Gujar | 08-May | 19A |
317 | Decent-BRM: Decentralization through Block Reward Mechanisms | Varul Srivastava and Sujit Gujar | 08-May | 20A |
321 | Fairness and Privacy Guarantees in Federated Contextual Bandits | Sambhav Solanki, Sujit Gujar and Shweta Jain | 08-May | 48A |
323 | Cognizing and Imitating Robotic Skills via a Dual Cognition-Action Architecture | Zixuan Chen, Ze Ji, Shuyang Liu, Jing Huo, Yiyu Chen and Yang Gao | 09-May | 11A |
335 | GOV-REK: Governed Reward Engineering Kernels for Designing Robust Multi-Agent Reinforcement Learning Systems | Ashish Rana, Michael Oesterle and Jannik Brinkmann | 08-May | 48B |
337 | Toward Socially Friendly Autonomous Driving Using Multi-agent Deep Reinforcement Learning | Jhih-Ching Yeh and Von-Wun Soo | 08-May | 4b |
343 | Persuasion by Shaping Beliefs about Multidimensional Features of a Thing | Kazunori Terada, Yasuo Noma and Masanori Hattori | 08-May | 4A |
348 | Competitive Analysis of Online Facility Open Problem | Binghan Wu, Wei Bao and Bing Zhou | 08-May | 31B |
351 | Guided Exploration in Reinforcement Learning via Monte Carlo Critic Optimization | Igor Kuznetsov | 08-May | 47B |
355 | JDRec: Practical Actor-Critic Framework for Online Combinatorial Recommender System | Xin Zhao, Jiaxin Li, Zhiwei Fang, Yuchen Guo, Jinyuan Zhao, Jie He, Wenlong Chen, Changping Peng and Guiguang Ding | 08-May | 46B |
356 | Reinforcement Nash Equilibrium Solver | Xinrun Wang, Chang Yang, Shuxin Li, Pengdeng Li, Xiao Huang, Hau Chan and Bo An | 09-May | 32A |
357 | Solving Offline 3D Bin Packing Problem with Large-sized Bin via Two-stage Deep Reinforcement Learning | Hao Yin, Fan Chen and Hongjie He | 08-May | 45B |
362 | Detecting Anomalous Agent Decision Sequences Based on Offline Imitation Learning | Chen Wang, Sarah Erfani, Tansu Alpcan and Christopher Leckie | 08-May | 44B |
366 | Explaining Sequences of Actions in Multi-agent Deep Reinforcement Learning Models | Khaing Phyo Wai, Minghong Geng, Shubham Pateria, Budhitama Subagdja and Ah-Hwee Tan | 08-May | 5A |
372 | Balanced and Incentivized Learning with Limited Shared Information in Multi-agent Multi-armed Bandit | Junning Shao, Siwei Wang and Zhixuan Fang | 09-May | 48B |
417 | Facility location games with task allocation | Zifan Gong, Minming Li and Houyu Zhou | 09-May | 24B |
418 | Sequential principal-agent problems with communication: efficient computation and learning | Jiarui Gan, Rupak Majumdar, Debmalya Mandal and Goran Radanovic | 09-May | 23B |
428 | Source Detection in Networks using the Stationary Distribution of a Markov Chain | Yael Sabato, Amos Azaria and Noam Hazon | 08-May | 5b |
433 | Near-Optimal Online Resource Allocation in the Random-Order Model | Saar Cohen and Noa Agmon | 08-May | 20B |
447 | Neurological Based Timing Mechanism for Reinforcement Learning | Michael Tarlton, Gustavo Mello and Anis Yazidi | 09-May | 47B |
462 | Inferring Lewisian common knowledge using theory of mind reasoning in a forward-chaining rule engine | Stephen Cranefield, Sriashalya Srivathsan and Jeremy Pitt | 08-May | 30B |
463 | Incentive-based MARL Approach for Commons Dilemmas in Property-based Environments | Lukasz Pelcner, Matheus Do Carmo Alves, Leandro Soriano Marcolino, Paula Harrison and Peter Atkinson | 08-May | 6b |
474 | Minimizing Negative Side Effects in Cooperative Multi-Agent Systems using Distributed Coordination | Moumita Choudhury, Sandhya Saisubramanian, Hao Zhang and Shlomo Zilberstein | 08-May | 29B |
478 | Optimal Task Assignment and Path Planning using Conflict-Based Search with Precedence and Temporal Constraints | Yu Quan Chong, Jiaoyang Li and Katia Sycara | 09-May | 31A |
479 | Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning | Zida Wu, Mathieu Lauriere, Samuel Jia Cong Chua, Matthieu Geist, Olivier Pietquin and Ankur Mehta | 09-May | 48A |
480 | Dual-Policy-Guided Offline Reinforcement Learning with Optimal Stopping | Weibo Jiang, Shaohui Li, Zhi Li, Yuxin Ke, Zhizhuo Jiang, Yaowen Li and Yu Liu | 09-May | 47A |
492 | Adaptive Discounting of Training Time Attacks | Ridhima Bector, Abhay Aradhya, Chai Quek and Zinovi Rabinovich | 09-May | 46A |
497 | vMFER: von Mises-Fisher Experience Resampling Based on Uncertainty of Gradient Directions for Policy Improvement of Actor-Critic Algorithms | Yiwen Zhu, Jinyi Liu, Wenya Wei, Qianyi Fu, Yujing Hu, Zhou Fang, Bo An, Jianye Hao, Tangjie Lv and Changjie Fan | 09-May | 45A |
510 | OPEx: A Large Language Model-Powered Framework for Embodied Instruction Following | Haochen Shi, Zhiyuan Sun, Xingdi Yuan, Marc-Alexandre Côté and Bang Liu | 09-May | 30A |
528 | ELA: Exploited Level Augmentation for Offline Learning in Zero-Sum Games | Shiqi Lei, Kanghoon Lee, Linjing Li, Jinkyoo Park and Jiachen Li | 09-May | 44A |
543 | HiMAP: Learning Heuristics-Informed Policies for Large-Scale Multi-Agent Pathfinding | Huijie Tang, Federico Berto, Zihan Ma, Chuanbo Hua, Kyuree Ahn and Jinkyoo Park | 08-May | 34A |
548 | Large Language Model Assissted Multi-Agent Dialogue for Ontology Alignment | Shiyao Zhang, Yuji Dong, Yichuan Zhang, Terry Payne and Jie Zhang | 09-May | 29A |
552 | Abstracting Assumptions in Structured Argumentation | Iosif Apostolakis, Zeynep G. Saribatur and Johannes P. Wallner | 09-May | 28A |
554 | Decentralized Competing Bandits in Many-to-One Matching Markets | Yirui Zhang and Zhixuan Fang | 09-May | 43A |
555 | Using Cooperative Game Theory to Prune Neural Networks | Mauricio Diaz-Ortiz Jr, Benjamin Kempinski, Daphne Cornelisse, Yoram Bachrach and Tal Kachman | 08-May | 19B |
559 | Bayesian Ensembles for Exploration in Deep Q-Learning | Pascal van der Vaart, Neil Yorke-Smith and Matthijs Spaan | 09-May | 42A |
561 | Understanding the impact of promotions on consumer behavior | Jarod Vanderlynden, Philippe Mathieu and Romain Warlop | 08-May | 7b |
563 | Electric Vehicle Routing for Emergency Power Supply with Deep Reinforcement Learning | Daisuke Kikuta, Hiroki Ikeuchi, Kengo Tajiri, Yuta Toyama, Masaki Nakamura and Yuusuke Nakano | 09-May | 1A |
567 | Metric Distortion Under Public-Spirited Voting | Amirreza Bagheridelouee, Marzie Nilipour, Masoud Seddighin and Maziar Shamsipour | 08-May | 18B |
569 | Factored MDP based Moving Target Defense with Dynamic Threat Modeling | Megha Bose, Praveen Paruchuri and Akshat Kumar | 08-May | 35A |
573 | Dual Role AoI-based Incentive Mechanism for HD map Crowdsourcing | Wentao Ye, Bo Liu, Yuan Luo and Jianwei Huang | 08-May | 17B |
574 | Fairness in Repeated House Allocation | Karl Jochen Micheel and Anaëlle Wilczynski | 08-May | 25A |
579 | Bootstrapped Policy Learning: Goal Shaping for Efficient Task-oriented Dialogue Policy Learning | Yangyang Zhao, Mehdi Dastani and Shihan Wang | 09-May | 41A |
583 | Unifying Regret and State-Action Space Coverage for Effective Unsupervised Environment Design | Jayden Teoh, Wenjun Li and Pradeep Varakantham | 09-May | 40B |
593 | Cournot Queueing Games with Applications to Mobility Systems | Matthew Sheldon, Dario Paccagnan and Giuliano Casale | 09-May | 22B |
595 | Verifying Proportionality in Temporal Multiwinner Voting | Edith Elkind, Svetlana Obraztsova and Nicholas Teh | 08-May | 26A |
599 | Combining Sentiment Analysis and Non-Bayesian Updating for Cooperative Decision-Making | Daniele Orner, Elizabeth Ondula, Nick Mumero and Richa Goyal | 09-May | 6A |
616 | Opinion Diffusion on Society Graphs Based on Approval Ballots | Jayakrishnan Madathil, Neeldhara Misra and Yash More | 09-May | 18B |
621 | On the Complexity of Candidates-Embedded Multiwinner Voting under the Hausdorff Function | Yongjie Yang | 09-May | 19B |
630 | HLG: Bridging Human Heuristic Knowledge and Deep Reinforcement Learning for Optimal Agent Performance | Bin Chen and Zehong Cao | 09-May | 39B |
637 | Reducing Systemic Risk in Financial Networks through Donations | Jinyun Tong, Bart De Keijzer and Carmine Ventre | 08-May | 9a |
658 | Strategic Cost Selection in Participatory Budgeting | Piotr Faliszewski, Łukasz Janeczko, Andrzej Kaczmarczyk, Grzegorz Lisowski, Piotr Skowron and Stanisław Szufa | 09-May | 20B |
659 | A Reinforcement Learning Framework For Studying Group And Individual Fairness | Alexandra Cimpean, Catholijn Jonker, Pieter Libin and Ann Nowé | 09-May | 38B |
672 | Unlocking the Potential of Machine Ethics with Explainability | Timo Speith | 08-May | 15A |
677 | BAR Nash Equilibrium and Application to Blockchain Design | Maxime Reynouard, Olga Gorelkina and Rida Laraki | 09-May | 21B |
696 | Addressing Permutation Challenges in Multi-Agent Reinforcement Learning | Somnath Hazra, Pallab Dasgupta and Soumyajit Dey | 09-May | 37B |
702 | Towards Zero Shot Learning in Restless Multi-armed Bandits | Yunfan Zhao, Nikhil Behari, Edward Hughes, Edwin Zhang, Dheeraj Nagaraj, Karl Tuyls, Aparna Taneja and Milind Tambe | 09-May | 36B |
708 | Computing Nash Equilibria in Multidimensional Congestion Games | Mohammad Irfan, Hau Chan and Jared Soundy | 09-May | 20B |
709 | Ontological modeling and reasoning for comparison and contrastive narration of robot plans | Alberto Olivares-Alarcos, Sergi Foix, Júlia Borràs, Gerard Canal and Guillem Alenyà | 09-May | 12A |
712 | Time-Constrained Restless Multi-Armed Bandits with Applications to City Service Scheduling | Yi Mao and Andrew Perrault | 09-May | 35B |
719 | Fuzzy Clustered Federated Learning Under Mixed Data Distributions | Peng Tang, Lifan Wang, Weidong Qiu, Zheng Huang and Qiangmin Wang | 08-May | 36A |
720 | Distribution of Chores with Information Asymmetry | Hadi Hosseini, Joshua Kavner, Tomasz Wąs and Lirong Xia | 09-May | 21B |
724 | ANOTO: Improving Automated Negotiation via Offline-to-Online Reinforcement Learning | Siqi Chen, Jianing Zhao, Kai Zhao, Gerhard Weiss, Fengyun Zhang, Ran Su, Yang Dong, Daqian Li and Kaiyou Lei | 09-May | 19B |
725 | The Selfishness Level of Social Dilemmas | Stefan Roesch, Stefanos Leonardos and Yali Du | 09-May | 18B |
727 | Consensus of Nonlinear Multi-Agent Systems with Semi-Markov Switching Under DoS Attacks | Sheng Tian, Hong Shen, Yuan Tian and Hui Tian | 09-May | 27A |
729 | Embracing Relational Reasoning in Multi-Agent Actor-Critic | Sharlin Utke, Jeremie Houssineau and Giovanni Montana | 09-May | 34B |
738 | Deep Learning for Population-Dependent Controls in Mean Field Control Problems with Common Noise | Gokce Dayanikli, Mathieu Lauriere and Jiacheng Zhang | 09-May | 46B |
744 | Psychophysiological Models of Cognitive States Can Be Operator-Agnostic | Erin Richardson, Savannah Buchner, Jacob Kintz, Torin Clark and Allison Anderson | 08-May | 6A |
749 | Mutual Information as Intrinsic Reward of Reinforcement Learning Agents for On-demand Ride Pooling | Xianjie Zhang, Jiahao Sun, Chen Gong, Kai Wang, Yifei Cao, Hao Chen and Yu Liu | 09-May | 2A |
757 | Which Games are Unaffected by Absolute Commitments? | Daji Landis and Nikolaj Ignatieff Schwartzbach | 09-May | 17B |
766 | MiKe: Task Scheduling for UAV-based Parcel Delivery | Viviana Arrigoni, Giulio Attenni, Novella Bartolini, Matteo Finelli and Gaia Maselli | 09-May | 3A |
768 | JaxMARL: Multi-Agent RL Environments in JAX | Alexander Rutherford, Benjamin Ellis, Matteo Gallici, Jonathan Cook, Andrei Lupu, Garðar Ingvarsson, Timon Willi, Akbir Khan, Christian Schroeder de Witt, Alexandra Souly, Saptarashmi Bandyopadhyay, Mikayel Samvelyan, Minqi Jiang, Robert Lange, Shimon Whiteson, Bruno Lacerda, Nick Hawes, Tim Rocktäschel, Chris Lu and Jakob Foerster | 09-May | 7A |
769 | Fairness and Cooperation between Independent Reinforcement Learners through Indirect Reciprocity | Jacobus Smit and Fernando Santos | 09-May | 33B |
773 | Approximately Fair Allocation of Indivisible Items with Random Valuations | Alessandro Aloisio, Vittorio Bilo, Antonio Mario Caruso, Michele Flammini and Cosimo Vinci | 09-May | 22B |
775 | Actual Trust in Multiagent Systems | Michael Akintunde, Vahid Yazdanpanah, Asieh Salehi Fathabadi, Corina Cirstea, Mehdi Dastani and Luc Moreau | 09-May | 26A |
776 | On General Epistemic Abstract Argumentation Frameworks | Gianvincenzo Alfano, Sergio Greco, Francesco Parisi and Irina Trubitsyna | 09-May | 25A |
790 | Fully Independent Communication in Multi-Agent Reinforcement Learning | Rafael Pina, Varuna De Silva, Corentin Artaud and Xiaolan Liu | 09-May | 40A |
796 | GLIDE-RL: Grounded Language Instruction through DEmonstration in RL | Chaitanya Kharyal, Sai Krishna Gottipati, Tanmay Sinha, Srijita Das and Matthew E. Taylor | 09-May | 39A |
805 | Towards Socially-Acceptable Multi-Criteria Resolution of the 4D-Contracts Repair Problem | Youssef Hamadi and Gauthier Picard | 09-May | 4A |
810 | Potential Games on Cubic Splines for Multi-Agent Motion Planning of Autonomous Agents | Sam Williams and Jyotirmoy Deshmukh | 09-May | 26B |
814 | Deceptive Path Planning via Reinforcement Learning with Graph Neural Networks | Michael Y Fatemi, Wesley A Suttle and Brian M Sadler | 09-May | 38A |
818 | Cutsets and EF1 Fair Division of Graphs | Jiehua Chen and William Zwicker | 09-May | 23B |
824 | Mechanism Design for Reducing Agent Distances to Prelocated Facilities | Hau Chan, Xinliang Fu, Minming Li and Chenhao Wang | 09-May | 17A |
840 | Taking Agent-Based Social Simulation to the Next Level Using Exascale Computing: Potential Use-Cases, Capacity Requirements and Threats. | Matt Hare, Douglas Salt, Ric Colasanti, Richard Milton, Mike Batty, Alison Heppenstall and Gary Polhill | 08-May | 10a |
842 | Distance-Aware Attentive Framework for Multi-Agent Collaborative Perception in Presence of Pose Error | Binyu Zhao, Wei Zhang and Zhaonian Zou | 09-May | 45B |
845 | Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning | Calarina Muslimani and Matthew Taylor | 09-May | 37A |
859 | Decision-Focused Model-based Reinforcement Learning for Reward Transfer | Abhishek Sharma, Sonali Parbhoo, Omer Gottesman and Finale Doshi-Velez | 09-May | 36A |
868 | Behaviour Modelling of Social Animals via Causal Structure Discovery and Graph Neural Networks | Gaël Gendron, Yang Chen, Mitchell Rogers, Yiping Liu, Mihailo Azhar, Shahrokh Heidari, David Arturo Soriano Valdez, Kobe Knowles, Padriac O’Leary, Simon Eyre, Michael Witbrock, Gillian Dobbie, Jiamou Liu and Patrice Delmas | 08-May | 11a |
876 | Geospatial Active Search for Preventing Evictions | Anindya Sarkar, Alex DiChristofano, Sanmay Das, Patrick Fowler, Nathan Jacobs and Yevgeniy Vorobeychik | 09-May | 5A |
880 | Efficient Size-based Hybrid Algorithm for Optimal Coalition Structure Generation | Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder and Tuomas Sandholm | 08-May | 16A |
894 | t-DGR: A Trajectory-Based Deep Generative Replay Method for Continual Learning in Decision Making | William Yue, Bo Liu and Peter Stone | 09-May | 44B |
901 | Decision Making in Non-Stationary Environments with Policy-Augmented Search | Ava Pettet, Yunuo Zhang, Baiting Luo, Kyle Wray, Hendrik Baier, Aron Laszka, Abhishek Dubey and Ayan Mukhopadhyay | 09-May | 43B |
904 | Anytime Multi-Agent Path Finding using Operator Parallelism in Large Neighborhood Search | Shao-Hung Chan, Zhe Chen, Dian-Lun Lin, Yue Zhang, Daniel Harabor, Sven Koenig, Tsung-Wei Huang and Thomy Phan | 09-May | 8A |
905 | On the existence of EFX under picky or non-differentiative agents | Maya Viswanathan and Ruta Mehta | 09-May | 24B |
914 | Fair Allocation of Conflicting Courses under Additive Utilities | Arpita Biswas, Yiduo Ke, Samir Khuller and Quanquan Liu | 09-May | 16B |
947 | A Specific-Purpose Linux Distribution for Embedded BDI-based Multi-agent Systems | Nilson Mori Lazarin, Carlos Pantoja and Jose Viterbo | 09-May | 8B |
969 | Proportionality and Free Riders: Committee Selection with Strategic Voters | Kerry Lu and Brandon Fain | 09-May | 15B |
974 | Defining Deception in Decision Making | Marwa Abdulhai, Micah Carroll, Justin Svegliato, Anca Dragan and Sergey Levine | 08-May | 7A |
976 | Quantifying Agent Interaction in Multi-agent Reinforcement Learning for Cost-efficient Generalization | Yuxin Chen, Chen Tang, Ran Tian, Chenran Li, Jinning Li, Masayoshi Tomizuka and Wei Zhan | 09-May | 35A |
982 | Efficiently Solving Min-Max Routing Problems via Parallel Autoregressive Policies | Federico Berto, Chuanbo Hua, Junyoung Park and Jinkyoo Park | 09-May | 34A |
983 | Bounding consideration probabilities in consider-then-choose ranking models | Ben Aoki-Sherwood, Catherine Bregou, David Liben-Nowell, Kiran Tomlinson and Thomas Zeng | 09-May | 42B |
997 | Optimal majority rules and quantitative Condorcet properties of setwise Kemeny voting schemes | Xuan Kien Phung and Sylvie Hamel | 09-May | 14B |
999 | Banzhaf Power in Hierarchical Games | John Randolph, Amy Greenwald and Denizalp Goktas | 09-May | 13B |
1010 | Efficient Collaboration with Unknown Agents: Ignoring Similar Agents without Checking Similarity | Yansong Li and Shuo Han | 09-May | 33A |
1015 | Difference of Convex Functions Programming for Policy Optimization in Reinforcement Learning | Akshat Kumar | 09-May | 32B |
1023 | A Multiagent Path Search Algorithm for Large-Scale Coalition Structure Generation | Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder and Tuomas Sandholm | 08-May | 16B |
1027 | On the Computational Complexity of Quasi-Variational Inequalities and Multi-Leader-Follower Games | Bruce M. Kapron and Koosha Samieefar | 09-May | 18A |
1029 | Non Stationary Bandits with Periodic Variation | Titas Chakraborty and Parth Shettiwar | 09-May | 31B |
1051 | Concurrency model of BDI programming frameworks: why should we control it? | Martina Baiardi, Samuele Burattini, Giovanni Ciatto, Danilo Pianini, Andrea Omicini and Alessandro Ricci | 09-May | 7B |
1067 | Reinforcement learning for question answering in programming domain using public community scoring as a human feedback | Alexey Gorbatovski and Sergey Kovalchuk | 08-May | 8A |
1068 | Improving Utilization and Sustainability of Low-power Wireless Sensors through Decentralized Role Allocation in a Multi-agent System | Ganesh Ramanathan, Simon Mayer, Simon Hess and Andres Gomez | 09-May | 6B |
1073 | Social Identities and Responsible Agency | Karthik Sama, Jayati Deshmukh and Srinath Srinivasa | 08-May | 12a |
1074 | Modelling the Dynamics of Subjective Identity in Allocation Games | Janvi Chhabra, Jayati Deshmukh and Srinath Srinivasa | 08-May | 2b |
1075 | Semantic Bridges in Engineering: Integrating Knowledge to Enable Autonomous Systems for Automation | Ganesh Ramanathan, Simon Mayer and Andrei Ciortea | 09-May | 5B |
1077 | User-centric Explanation Strategies for Interactive Recommenders | Berk Buzcu, Emre Kuru and Reyhan Aydogan | 08-May | 8B |
1083 | Who gets the Maximal Extractable Value? A Dynamic Sharing Blockchain Mechanism | Georgios Chionas, Pedro Braga, Stefanos Leonardos, Carmine Ventre, Georgios Piliouras and Piotr Krysta | 09-May | 41B |
1095 | Quantum Circuit Design: A Reinforcement Learning Challenge | Philipp Altmann, Adelina Bärligea, Jonas Stein, Michael Kölle, Thomas Gabor, Thomy Phan and Claudia Linnhof-Popien | 09-May | 30B |
1097 | Simple $k$-crashing Plan with a Good Approximation Ratio | Ruixi Luo, Kai Jin and Zelin Ye | 09-May | 27B |
1100 | Risk-Sensitive Multi-Agent Reinforcement Learning in Network Aggregative Markov Games | Hafez Ghaemi, Hamed Kebriaei, Alireza Ramezani Moghaddam and Majid Nili Ahmadabadi | 09-May | 29B |
1107 | Truthful and Stable One-sided Matching on Networks | Tianyi Yang, Yuxiang Zhai, Dengji Zhao, Xinwei Song and Miao Li | 09-May | 12B |
1115 | Deep Hawkes Process for High-Frequency Market Making | Pankaj Kumar | 08-May | 3B |
1116 | Agent-Based Triangle Counting and its Applications in Anonymous Graphs | Prabhat Kumar Chand, Apurba Das and Anisur Rahaman Molla | 09-May | 13A |
1124 | Aleatoric Predicates: Reasoning about Marbles | Tim French | 09-May | 28B |
1128 | Hybrid Participatory Budgeting: Divisible, Indivisible, and Beyond | Gogulapati Sreedurga | 09-May | 11B |