ID |
Authors |
Title |
3 |
Kenjiro Takazawa |
Pure Nash Equilibria in Weighted Congestion Games with Complementarities and Beyond |
8 |
David Milec, Ondřej Kubíček and Viliam Lisy |
Continual Depth-limited Responses for Computing Counter-strategies in Sequential Games |
12 |
Haipeng Zhang, Zhiwen Wang and Na Li |
MATLight: Traffic Signal Coordinated Control Algorithm Based on Heterogeneous-Agent Mirror Learning With Transformer |
13 |
Ben Armstrong and Kate Larson |
Liquid Democracy for Low-Cost Ensemble Pruning |
14 |
Abhijat Biswas, Badal Arun Pardhi, Caleb Chuck, Jarrett Holtz, Scott Niekum, Henny Admoni and Alessandro Allievi |
Gaze Supervision for Mitigating Causal Confusion in Driving Agents |
21 |
Ram Rachum, Yonatan Nakar, Bill Tomlinson, Nitay Alon and Reuth Mirsky |
Emergent Dominance Hierarchies in Reinforcement Learning Agents |
32 |
Mathieu Mari, Michał Pawłowski, Runtian Ren and Piotr Sankowski |
Multi-level aggregation with delays and stochastic arrivals |
41 |
Alexander W. Goodall and Francesco Belardinelli |
Leveraging Approximate Model-based Shielding for Probabilistic Safety Guarantees in Continuous Environments |
50 |
Kazi Ashik Islam, Da Qi Chen, Madhav Marathe, Henning Mortveit, Samarth Swarup and Anil Vullikanti |
Strategic Routing and Scheduling for Evacuations |
67 |
Daniel Melcer, Christopher Amato and Stavros Tripakis |
Shield Decentralization for Safe Reinforcement Learning in General Partially Observable Multi-Agent Environments |
77 |
Thayne T. Walker, Nathan Sturtevant and Ariel Felner |
Clique Analysis and Bypassing in Continuous-Time Conflict-Based Search |
85 |
Jiafei Lyu, Le Wan, Xiu Li and Zongqing Lu |
Towards Understanding How to Reduce Generalization Gap in Visual Reinforcement Learning |
94 |
Dave de Jonge and Laura Rodriguez Cima |
Attila: a Negotiating Agent for the Game of Diplomacy, Based on Purely Symbolic A.I. |
107 |
Dapeng Li, Zhiwei Xu, Bin Zhang, Guangchong Zhou, Zeren Zhang and Guoliang Fan |
From Explicit Communication to Tacit Cooperation: A Novel Paradigm for Cooperative MARL |
108 |
Pankaj Deoli, Rohit Kumar, Axel Vierling and Karsten Berns |
Evaluation of Robustness of Off-Road Autonomous Driving Segmentation against Adversarial Attacks: A Dataset-Centric Study |
113 |
Glareh Mir and Michael Beetz |
Simulated Robotic Soft Body Manipulation |
114 |
Yu Niu, Hengxu Zhao and Lei Yu |
MA-MIX: Value Function Decomposition for Cooperative Multiagent Reinforcement Learning Based on Multi-Head Attention Mechanism |
115 |
Ayşe Mutlu Derya |
A Comparison of the Myerson Value and the Position Value |
117 |
Tamara C.P. Florijn, Pinar Yolum and Tim Baarslag |
A Negotiator’s Backup Plan: Optimal Concessions with a Reservation Value |
140 |
Jérôme Botoko Ekila, Jens Nevens, Lara Verheyen, Katrien Beuls and Paul Van Eecke |
Decentralised Emergence of Robust and Adaptive Linguistic Conventions in Populations of Autonomous Agents Grounded in Continuous Worlds |
142 |
Minghong Geng, Shubham Pateria, Budhitama Subagdja and Ah-Hwee Tan |
Benchmarking MARL on Long Horizon Sequential Multi-Objective Tasks |
143 |
Yatharth Kumar, Sarfaraz Equbal, Rohit Gurjar, Swaprava Nath and Rohit Vaish |
Fair Scheduling of Indivisible Chores |
152 |
Qitong Kang, Fuyong Wang, Zhongxin Liu and Zengqiang Chen |
TIMAT: Temporal Information Multi-Agent Transformer |
154 |
Huihui Zhang |
Bellman Momentum on Deep Reinforcement Learning |
158 |
Yasushi Kawase, Bodhayan Roy and Mohammad Azharuddin Sanpui |
Contiguous Allocation of Binary Valued Indivisible Items on a Path |
166 |
Nicolas Bessone, Payam Zahadat and Kasper Stoy |
Decentralized Control of Distributed Manipulators: An Information Diffusion Approach |
175 |
Ramsundar Anandanarayanan, Swaprava Nath and Rohit Vaish |
Charging Electric Vehicles Fairly and Efficiently |
180 |
Kaifeng Zhang, Rui Zhao, Ziming Zhang and Yang Gao |
Auto-Encoding Adversarial Imitation Learning |
190 |
Mihail Stojanovski, Nadjet Bourdache, Grégory Bonnet and Mouaddib Abdel-Illah |
Ethical Markov Decision Processes with Moral Worth as Rewards |
197 |
Emanuel Tewolde and Vincent Conitzer |
Game Transformations That Preserve Nash Equilibria or Best Response Sets |
213 |
Yihong Chen, Cong Wang, Tianpei Yang, Meng Wang, Yingfeng Chen, Jifei Zhou, Chaoyi Zhao, Xinfeng Zhang, Zeng Zhao, Changjie Fan, Zhipeng Hu, Rong Xiong and Long Zeng |
Mastering Robot Control through Point-based Reinforcement Learning with Pre-training |
215 |
Chenxu Wang, Zilong Chen and Huaping Liu |
On the Utility of External Agent Intention Predictor for Human-AI Coordination |
219 |
Jean Marie Lagniez, Emmanuel Lonca and Jean-Guy Mailly |
A SAT-based Approach for Argumentation Dynamics |
221 |
Yao Zhang, Shanshan Zheng and Dengji Zhao |
Optimal Diffusion Auctions |
222 |
Erwan Escudie, Laetitia Matignon and Jacques Saraydaryan |
Attention Graph for Multi-Robot Social Navigation with Deep Reinforcement Learning |
225 |
Pranavi Pathakota, Hardik Meisheri and Harshad Khadilkar |
DCT: Dual Channel Training of Action Embeddings for Reinforcement Learning with Large Discrete Action Spaces |
226 |
Wenlong Wang and Thomas Pfeiffer |
Decision Market Based Learning For Multi-agent Contextual Bandit Problems |
235 |
Kai Zhao, Jianye Hao, Yi Ma, Jinyi Liu, Yan Zheng and Zhaopeng Meng |
ENOTO: Improving Offline-to-Online Reinforcement Learning with Q-Ensembles |
238 |
Emre Erdogan, Rineke Verbrugge and Pinar Yolum |
Computational Theory of Mind with Abstractions for Effective Human-Agent Collaboration |
239 |
Pedro P. Santos, Diogo Carvalho, Miguel Vasco, Alberto Sardinha, Pedro A. Santos, Ana Paiva and Francisco Melo |
Centralized Training with Hybrid Execution in Multi-Agent Reinforcement Learning |
241 |
Márton Benedek, Péter Biró, Gergely Csáji, Matthew Johnson, Daniël Paulusma and Xin Ye |
Computing Balanced Solutions for Large International Kidney Exchange Schemes When Cycle Length Is Unbounded |
249 |
Yongsheng Mei, Hanhan Zhou and Tian Lan |
Projection-Optimal Monotonic Value Function Factorization in Multi-Agent Reinforcement Learning |
255 |
Maxime Toquebiau, Nicolas Bredeche, Faïz Ben Amar and Jae-Yun Jun |
Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent Deep Reinforcement Learning |
256 |
Saad Khan, Mayank Baranwal and Srikant Sukumar |
Decentralized Safe Control for Multi-Robot Navigation in Dynamic Environments with Limited Sensing |
257 |
Rustam Galimullin and Louwe B. Kuijer |
Synthesizing social laws with ATL conditions |
261 |
Sai Srivatsa Ravindranath, Zhe Feng, Shira Li, Jonathan Ma, Scott Kominers and David Parkes |
Deep Learning for Two-Sided Matching Markets |
262 |
Tesfay Zemuy Gebrekidan, Sebastian Stein and Timothy Norman |
Combinatorial Client-Master Multiagent Deep Reinforcement Learning for Task Offloading in Mobile Edge Computing |
268 |
Alexander Mendelsohn, Donald Sofge and Michael Otte |
Enhancing Search and Rescue Capabilities in Hazardous Communication-Denied Environments through Path-Based Sensors with Backtracking |
272 |
Ayhan Alp Aydeniz, Enrico Marchesini, Christopher Amato and Kagan Tumer |
Entropy Seeking Constrained Multiagent Reinforcement Learning |
283 |
Everardo Gonzalez, Siddarth Viswanathan and Kagan Tumer |
Indirect Credit Assignment in a Multiagent System |
290 |
Andrew Festa, Gaurav Dixit and Kagan Tumer |
Influence-Focused Asymmetric Island Model |
298 |
Hao Zhang, Tianpei Yang, Yan Zheng, Jianye Hao and Matthew E. Taylor |
PADDLE: Logic Program Guided Policy Reuse in Deep Reinforcement Learning |
301 |
Joanna Kaczmarek and Jörg Rothe |
NP^PP-Completeness of Control by Adding Players to Change the Penrose–Banzhaf Power Index in Weighted Voting Games |
305 |
Archit Sood, Shweta Jain and Sujit Gujar |
Fairness of Exposure in Online Restless Multi-armed Bandits |
307 |
Sankarshan Damle and Sujit Gujar |
Analyzing Crowdfunding of Public Projects Under Dynamic Beliefs |
312 |
Sankarshan Damle, Varul Srivastava and Sujit Gujar |
No Transaction Fees? No Problem! Achieving Fairness in Transaction Fee Mechanism Design |
317 |
Varul Srivastava and Sujit Gujar |
Decent-BRM: Decentralization through Block Reward Mechanisms |
321 |
Sambhav Solanki, Sujit Gujar and Shweta Jain |
Fairness and Privacy Guarantees in Federated Contextual Bandits |
323 |
Zixuan Chen, Ze Ji, Shuyang Liu, Jing Huo, Yiyu Chen and Yang Gao |
Cognizing and Imitating Robotic Skills via a Dual Cognition-Action Architecture |
335 |
Ashish Rana, Michael Oesterle and Jannik Brinkmann |
GOV-REK: Governed Reward Engineering Kernels for Designing Robust Multi-Agent Reinforcement Learning Systems |
337 |
Jhih-Ching Yeh and Von-Wun Soo |
Toward Socially Friendly Autonomous Driving Using Multi-agent Deep Reinforcement Learning |
343 |
Kazunori Terada, Yasuo Noma and Masanori Hattori |
Persuasion by Shaping Beliefs about Multidimensional Features of a Thing |
348 |
Binghan Wu, Wei Bao and Bing Zhou |
Competitive Analysis of Online Facility Open Problem |
351 |
Igor Kuznetsov |
Guided Exploration in Reinforcement Learning via Monte Carlo Critic Optimization |
355 |
Xin Zhao, Jiaxin Li, Zhiwei Fang, Yuchen Guo, Jinyuan Zhao, Jie He, Wenlong Chen, Changping Peng and Guiguang Ding |
JDRec: Practical Actor-Critic Framework for Online Combinatorial Recommender System |
356 |
Xinrun Wang, Chang Yang, Shuxin Li, Pengdeng Li, Xiao Huang, Hau Chan and Bo An |
Reinforcement Nash Equilibrium Solver |
357 |
Hao Yin, Fan Chen and Hongjie He |
Solving Offline 3D Bin Packing Problem with Large-sized Bin via Two-stage Deep Reinforcement Learning |
362 |
Chen Wang, Sarah Erfani, Tansu Alpcan and Christopher Leckie |
Detecting Anomalous Agent Decision Sequences Based on Offline Imitation Learning |
366 |
Khaing Phyo Wai, Minghong Geng, Shubham Pateria, Budhitama Subagdja and Ah-Hwee Tan |
Explaining Sequences of Actions in Multi-agent Deep Reinforcement Learning Models |
372 |
Junning Shao, Siwei Wang and Zhixuan Fang |
Balanced and Incentivized Learning with Limited Shared Information in Multi-agent Multi-armed Bandit |
417 |
Zifan Gong, Minming Li and Houyu Zhou |
Facility location games with task allocation |
418 |
Jiarui Gan, Rupak Majumdar, Debmalya Mandal and Goran Radanovic |
Sequential principal-agent problems with communication: efficient computation and learning |
428 |
Yael Sabato, Amos Azaria and Noam Hazon |
Source Detection in Networks using the Stationary Distribution of a Markov Chain |
433 |
Saar Cohen and Noa Agmon |
Near-Optimal Online Resource Allocation in the Random-Order Model |
447 |
Michael Tarlton, Gustavo Mello and Anis Yazidi |
Neurological Based Timing Mechanism for Reinforcement Learning |
462 |
Stephen Cranefield, Sriashalya Srivathsan and Jeremy Pitt |
Inferring Lewisian common knowledge using theory of mind reasoning in a forward-chaining rule engine |
463 |
Lukasz Pelcner, Matheus Do Carmo Alves, Leandro Soriano Marcolino, Paula Harrison and Peter Atkinson |
Incentive-based MARL Approach for Commons Dilemmas in Property-based Environments |
474 |
Moumita Choudhury, Sandhya Saisubramanian, Hao Zhang and Shlomo Zilberstein |
Minimizing Negative Side Effects in Cooperative Multi-Agent Systems using Distributed Coordination |
478 |
Yu Quan Chong, Jiaoyang Li and Katia Sycara |
Optimal Task Assignment and Path Planning using Conflict-Based Search with Precedence and Temporal Constraints |
479 |
Zida Wu, Mathieu Lauriere, Samuel Jia Cong Chua, Matthieu Geist, Olivier Pietquin and Ankur Mehta |
Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning |
480 |
Weibo Jiang, Shaohui Li, Zhi Li, Yuxin Ke, Zhizhuo Jiang, Yaowen Li and Yu Liu |
Dual-Policy-Guided Offline Reinforcement Learning with Optimal Stopping |
492 |
Ridhima Bector, Abhay Aradhya, Chai Quek and Zinovi Rabinovich |
Adaptive Discounting of Training Time Attacks |
497 |
Yiwen Zhu, Jinyi Liu, Wenya Wei, Qianyi Fu, Yujing Hu, Zhou Fang, Bo An, Jianye Hao, Tangjie Lv and Changjie Fan |
vMFER: von Mises-Fisher Experience Resampling Based on Uncertainty of Gradient Directions for Policy Improvement of Actor-Critic Algorithms |
510 |
Haochen Shi, Zhiyuan Sun, Xingdi Yuan, Marc-Alexandre Côté and Bang Liu |
OPEx: A Large Language Model-Powered Framework for Embodied Instruction Following |
528 |
Shiqi Lei, Kanghoon Lee, Linjing Li, Jinkyoo Park and Jiachen Li |
ELA: Exploited Level Augmentation for Offline Learning in Zero-Sum Games |
543 |
Huijie Tang, Federico Berto, Zihan Ma, Chuanbo Hua, Kyuree Ahn and Jinkyoo Park |
HiMAP: Learning Heuristics-Informed Policies for Large-Scale Multi-Agent Pathfinding |
548 |
Shiyao Zhang, Yuji Dong, Yichuan Zhang, Terry Payne and Jie Zhang |
Large Language Model Assissted Multi-Agent Dialogue for Ontology Alignment |
552 |
Iosif Apostolakis, Zeynep G. Saribatur and Johannes P. Wallner |
Abstracting Assumptions in Structured Argumentation |
554 |
Yirui Zhang and Zhixuan Fang |
Decentralized Competing Bandits in Many-to-One Matching Markets |
555 |
Mauricio Diaz-Ortiz Jr, Benjamin Kempinski, Daphne Cornelisse, Yoram Bachrach and Tal Kachman |
Pruning Neural Networks Using Cooperative Game Theory |
559 |
Pascal van der Vaart, Neil Yorke-Smith and Matthijs Spaan |
Bayesian Ensembles for Exploration in Deep Q-Learning |
561 |
Jarod Vanderlynden, Philippe Mathieu and Romain Warlop |
Understanding the impact of promotions on consumer behavior |
563 |
Daisuke Kikuta, Hiroki Ikeuchi, Kengo Tajiri, Yuta Toyama, Masaki Nakamura and Yuusuke Nakano |
Electric Vehicle Routing for Emergency Power Supply with Deep Reinforcement Learning |
567 |
Amirreza Bagheridelouee, Marzie Nilipour, Masoud Seddighin and Maziar Shamsipour |
Metric Distortion Under Public-Spirited Voting |
569 |
Megha Bose, Praveen Paruchuri and Akshat Kumar |
Factored MDP based Moving Target Defense with Dynamic Threat Modeling |
573 |
Wentao Ye, Bo Liu, Yuan Luo and Jianwei Huang |
Dual Role AoI-based Incentive Mechanism for HD map Crowdsourcing |
574 |
Karl Jochen Micheel and Anaëlle Wilczynski |
Fairness in Repeated House Allocation |
579 |
Yangyang Zhao, Mehdi Dastani and Shihan Wang |
Bootstrapped Policy Learning: Goal Shaping for Efficient Task-oriented Dialogue Policy Learning |
583 |
Jayden Teoh, Wenjun Li and Pradeep Varakantham |
Unifying Regret and State-Action Space Coverage for Effective Unsupervised Environment Design |
593 |
Matthew Sheldon, Dario Paccagnan and Giuliano Casale |
Cournot Games for Closed Cournot Queueing Games with Applications to Mobility Systems Networks |
595 |
Edith Elkind, Svetlana Obraztsova and Nicholas Teh |
Verifying Proportionality in Temporal Voting |
599 |
Daniele Orner, Elizabeth Ondula, Nick Mumero and Richa Goyal |
Sentimental Agents: Combining Sentiment Analysis and Non-Bayesian Updating for Cooperative Decision-Making |
616 |
Jayakrishnan Madathil, Neeldhara Misra and Yash More |
Opinion Diffusion on Society Graphs Based on Approval Ballots |
621 |
Yongjie Yang |
On the Complexity of Candidates-Embedded Multiwinner Voting under the Hausdorff Function |
630 |
Bin Chen and Zehong Cao |
HLG: Bridging Human Heuristic Knowledge and Deep Reinforcement Learning for Optimal Agent Performance |
637 |
Jinyun Tong, Bart De Keijzer and Carmine Ventre |
Reducing Systemic Risk in Financial Networks through Donations |
658 |
Piotr Faliszewski, Łukasz Janeczko, Andrzej Kaczmarczyk, Grzegorz Lisowski, Piotr Skowron and Stanisław Szufa |
Strategic Cost Selection in Participatory Budgeting |
659 |
Alexandra Cimpean, Catholijn Jonker, Pieter Libin and Ann Nowé |
A Reinforcement Learning Framework For Studying Group And Individual Fairness |
672 |
Timo Speith |
Unlocking the Potential of Machine Ethics with Explainability |
677 |
Maxime Reynouard, Olga Gorelkina and Rida Laraki |
BAR Nash Equilibrium and Application to Blockchain Design |
696 |
Somnath Hazra, Pallab Dasgupta and Soumyajit Dey |
Addressing Permutation Challenges in Multi-Agent Reinforcement Learning |
702 |
Yunfan Zhao, Nikhil Behari, Edward Hughes, Edwin Zhang, Dheeraj Nagaraj, Karl Tuyls, Aparna Taneja and Milind Tambe |
Towards Zero Shot Learning in Restless Multi-armed Bandits |
708 |
Mohammad Irfan, Hau Chan and Jared Soundy |
Computing Nash Equilibria in Multidimensional Congestion Games |
709 |
Alberto Olivares-Alarcos, Sergi Foix, Júlia Borràs, Gerard Canal and Guillem Alenyà |
Ontological modeling and reasoning for comparison and contrastive narration of robot plans |
712 |
Yi Mao and Andrew Perrault |
Time-Constrained Restless Multi-Armed Bandits with Applications to City Service Scheduling |
719 |
Peng Tang, Lifan Wang, Weidong Qiu, Zheng Huang and Qiangmin Wang |
Fuzzy Clustered Federated Learning Under Mixed Data Distributions |
720 |
Hadi Hosseini, Joshua Kavner, Tomasz Wąs and Lirong Xia |
Distribution of Chores with Information Asymmetry |
724 |
Siqi Chen, Jianing Zhao, Kai Zhao, Gerhard Weiss, Fengyun Zhang, Ran Su, Yang Dong, Daqian Li and Kaiyou Lei |
ANOTO: Improving Automated Negotiation via Offline-to-Online Reinforcement Learning |
725 |
Stefan Roesch, Stefanos Leonardos and Yali Du |
The Selfishness Level of Social Dilemmas |
727 |
Sheng Tian, Hong Shen, Yuan Tian and Hui Tian |
Consensus of Nonlinear Multi-Agent Systems with Semi-Markov Switching Under DoS Attacks |
729 |
Sharlin Utke, Jeremie Houssineau and Giovanni Montana |
Embracing Relational Reasoning in Multi-Agent Actor-Critic |
738 |
Gokce Dayanikli, Mathieu Lauriere and Jiacheng Zhang |
Deep Learning for Population-Dependent Controls in Mean Field Control Problems with Common Noise |
744 |
Erin Richardson, Savannah Buchner, Jacob Kintz, Torin Clark and Allison Anderson |
Psychophysiological Models of Cognitive States Can Be Operator-Agnostic |
749 |
Xianjie Zhang, Jiahao Sun, Chen Gong, Kai Wang, Yifei Cao, Hao Chen and Yu Liu |
Mutual Information as Intrinsic Reward of Reinforcement Learning Agents for On-demand Ride Pooling |
757 |
Daji Landis and Nikolaj Ignatieff Schwartzbach |
Which Games are Unaffected by Absolute Commitments? |
766 |
Viviana Arrigoni, Giulio Attenni, Novella Bartolini, Matteo Finelli and Gaia Maselli |
MiKe: Task Scheduling for UAV-based Parcel Delivery |
768 |
Alexander Rutherford, Benjamin Ellis, Matteo Gallici, Jonathan Cook, Andrei Lupu, Garðar Ingvarsson, Timon Willi, Akbir Khan, Christian Schroeder de Witt, Alexandra Souly, Saptarashmi Bandyopadhyay, Mikayel Samvelyan, Minqi Jiang, Robert Lange, Shimon Whiteson, Bruno Lacerda, Nick Hawes, Tim Rocktäschel, Chris Lu and Jakob Foerster |
JaxMARL: Multi-Agent RL Environments in JAX |
769 |
Jacobus Smit and Fernando Santos |
Fairness and Cooperation between Independent Reinforcement Learners through Indirect Reciprocity |
773 |
Alessandro Aloisio, Vittorio Bilo, Antonio Mario Caruso, Michele Flammini and Cosimo Vinci |
Approximately Fair Allocation of Indivisible Items with Random Valuations |
775 |
Michael Akintunde, Vahid Yazdanpanah, Asieh Salehi Fathabadi, Corina Cirstea, Mehdi Dastani and Luc Moreau |
Actual Trust in Multiagent Systems |
776 |
Gianvincenzo Alfano, Sergio Greco, Francesco Parisi and Irina Trubitsyna |
General Epistemic Abstract Argumentation Framework: Semantics and Complexity |
790 |
Rafael Pina, Varuna De Silva, Corentin Artaud and Xiaolan Liu |
Fully Independent Communication in Multi-Agent Reinforcement Learning |
796 |
Chaitanya Kharyal, Sai Krishna Gottipati, Tanmay Sinha, Srijita Das and Matthew E. Taylor |
GLIDE-RL: Grounded Language Instruction through DEmonstration in RL |
805 |
Youssef Hamadi and Gauthier Picard |
Towards Socially-Acceptable Multi-Criteria Resolution of the 4D-Contracts Repair Problem |
810 |
Sam Williams and Jyotirmoy Deshmukh |
Potential Games on Cubic Splines for Multi-Agent Motion Planning of Autonomous Agents |
814 |
Michael Y Fatemi, Wesley A Suttle and Brian M Sadler |
Deceptive Path Planning via Reinforcement Learning with Graph Neural Networks |
818 |
Jiehua Chen and William Zwicker |
Cutsets and EF1 Fair Division of Graphs |
824 |
Hau Chan, Xinliang Fu, Minming Li and Chenhao Wang |
Mechanism Design for Reducing Agent Distances to Prelocated Facilities |
840 |
Matt Hare, Douglas Salt, Ric Colasanti, Richard Milton, Mike Batty, Alison Heppenstall and Gary Polhill |
Taking Agent-Based Social Simulation to the Next Level Using Exascale Computing: Potential Use-Cases, Capacity Requirements and Threats. |
842 |
Binyu Zhao, Wei Zhang and Zhaonian Zou |
Distance-Aware Attentive Framework for Multi-Agent Collaborative Perception in Presence of Pose Error |
845 |
Calarina Muslimani and Matthew Taylor |
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning |
868 |
Gaël Gendron, Yang Chen, Mitchell Rogers, Yiping Liu, Mihailo Azhar, Shahrokh Heidari, David Arturo Soriano Valdez, Kobe Knowles, Padriac O’Leary, Simon Eyre, Michael Witbrock, Gillian Dobbie, Jiamou Liu and Patrice Delmas |
Behaviour Modelling of Social Animals via Causal Structure Discovery and Graph Neural Networks |
876 |
Anindya Sarkar, Alex DiChristofano, Sanmay Das, Patrick Fowler, Nathan Jacobs and Yevgeniy Vorobeychik |
Geospatial Active Search for Preventing Evictions |
880 |
Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder and Tuomas Sandholm |
Efficient Size-based Hybrid Algorithm for Optimal Coalition Structure Generation |
894 |
William Yue, Bo Liu and Peter Stone |
Overview of t-DGR: A Trajectory-Based Deep Generative Replay Method for Continual Learning in Decision Making |
901 |
Ava Pettet, Yunuo Zhang, Baiting Luo, Kyle Wray, Hendrik Baier, Aron Laszka, Abhishek Dubey and Ayan Mukhopadhyay |
Decision Making in Non-Stationary Environments with Policy-Augmented Search |
904 |
Shao-Hung Chan, Zhe Chen, Dian-Lun Lin, Yue Zhang, Daniel Harabor, Sven Koenig, Tsung-Wei Huang and Thomy Phan |
Anytime Multi-Agent Path Finding using Operator Parallelism in Large Neighborhood Search |
905 |
Maya Viswanathan and Ruta Mehta |
On the existence of EFX under picky or non-differentiative agents |
914 |
Arpita Biswas, Yiduo Ke, Samir Khuller and Quanquan Liu |
Fair Allocation of Conflicting Courses under Additive Utilities |
947 |
Nilson Mori Lazarin, Carlos Pantoja and Jose Viterbo |
A Specific-Purpose Linux Distribution for Embedded BDI-based Multi-agent Systems |
974 |
Marwa Abdulhai, Micah Carroll, Justin Svegliato, Anca Dragan and Sergey Levine |
Defining Deception in Decision Making |
976 |
Yuxin Chen, Chen Tang, Ran Tian, Chenran Li, Jinning Li, Masayoshi Tomizuka and Wei Zhan |
Quantifying Agent Interaction in Multi-agent Reinforcement Learning for Cost-efficient Generalization |
982 |
Federico Berto, Chuanbo Hua, Junyoung Park and Jinkyoo Park |
Efficiently Solving Min-Max Routing Problems via Parallel Autoregressive Policies |
983 |
Ben Aoki-Sherwood, Catherine Bregou, David Liben-Nowell, Kiran Tomlinson and Thomas Zeng |
Bounding consideration probabilities in consider-then-choose ranking models |
997 |
Xuan Kien Phung and Sylvie Hamel |
Optimal majority rules and quantitative Condorcet properties of setwise Kemeny voting schemes |
999 |
John Randolph, Amy Greenwald and Denizalp Goktas |
Banzhaf Power in Hierarchical Games |
1010 |
Yansong Li and Shuo Han |
Efficient Collaboration with Unknown Agents: Ignoring Similar Agents without Checking Similarity |
1015 |
Akshat Kumar |
Difference of Convex Functions Programming for Policy Optimization in Reinforcement Learning |
1023 |
Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder and Tuomas Sandholm |
A Multiagent Path Search Algorithm for Large-Scale Coalition Structure Generation |
1027 |
Bruce M. Kapron and Koosha Samieefar |
On the Computational Complexity of Quasi-Variational Inequalities and Multi-Leader-Follower Games |
1029 |
Titas Chakraborty and Parth Shettiwar |
Non Stationary Bandits with Periodic Variation |
1051 |
Martina Baiardi, Samuele Burattini, Giovanni Ciatto, Danilo Pianini, Andrea Omicini and Alessandro Ricci |
Concurrency model of BDI programming frameworks: why should we control it? |
1067 |
Alexey Gorbatovski and Sergey Kovalchuk |
Reinforcement learning for question answering in programming domain using public community scoring as a human feedback |
1068 |
Ganesh Ramanathan, Simon Mayer, Simon Hess and Andres Gomez |
Improving Utilization and Sustainability of Low-power Wireless Sensors through Decentralized Role Allocation in a Multi-agent System |
1073 |
Karthik Sama, Jayati Deshmukh and Srinath Srinivasa |
Social Identities and Responsible Agency |
1074 |
Janvi Chhabra, Jayati Deshmukh and Srinath Srinivasa |
Modelling the Dynamics of Subjective Identity in Allocation Games |
1075 |
Ganesh Ramanathan, Simon Mayer and Andrei Ciortea |
Semantic Bridges in Engineering: Integrating Knowledge to Enable Autonomous Systems for Automation |
1077 |
Berk Buzcu, Emre Kuru and Reyhan Aydogan |
User-centric Explanation Strategies for Interactive Recommenders |
1083 |
Georgios Chionas, Pedro Braga, Stefanos Leonardos, Carmine Ventre, Georgios Piliouras and Piotr Krysta |
Who gets the Maximal Extractable Value? A Dynamic Sharing Blockchain Mechanism |
1095 |
Philipp Altmann, Adelina Bärligea, Jonas Stein, Michael Kölle, Thomas Gabor, Thomy Phan and Claudia Linnhof-Popien |
Challenges for Reinforcement Learning in Quantum Computing |
1097 |
Ruixi Luo, Kai Jin and Zelin Ye |
Simple $k$-crashing Plan with a Good Approximation Ratio |
1100 |
Hafez Ghaemi, Hamed Kebriaei, Alireza Ramezani Moghaddam and Majid Nili Ahmadabadi |
Risk-Sensitive Multi-Agent Reinforcement Learning in Network Aggregative Markov Games |
1107 |
Tianyi Yang, Yuxiang Zhai, Dengji Zhao, Xinwei Song and Miao Li |
Truthful and Stable One-sided Matching on Networks |
1115 |
Pankaj Kumar |
Deep Hawkes Process for High-Frequency Market Making |
1116 |
Prabhat Kumar Chand, Apurba Das and Anisur Rahaman Molla |
Agent-Based Triangle Counting and its Applications in Anonymous Graphs |
1124 |
Tim French |
Aleatoric Predicates: Reasoning about Marbles |
1128 |
Gogulapati Sreedurga |
Hybrid Participatory Budgeting: Divisible, Indivisible, and Beyond |