default search action
24th AAMAS 2025: Detroit, MI, USA
- Sanmay Das, Ann Nowé, Yevgeniy Vorobeychik:
Proceedings of the 24th International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2025, Detroit, MI, USA, May 19-23, 2025. International Foundation for Autonomous Agents and Multiagent Systems / ACM 2025, ISBN 979-8-4007-1426-9
Keynote Talks
- Virginia Dignum:
Responsible AI and Autonomous Agents: Governance, Ethics, and Sustainable Innovation. 1-2 - Rada Mihalcea:
The Agent Paradox: Can Multi-Agent Systems Replicate the Complexity of Human Cognition and Social Behavior? 3 - Jeffrey S. Rosenschein:
Multiagent Systems, and the Search for Appropriate Foundations: A Personal Journey and Retrospective. 4
Research Paper Track
- Amir Abolfazli, Zekun Song, Avishek Anand, Wolfgang Nejdl:
Enhancing Offline Reinforcement Learning with Curriculum Learning-Based Trajectory Valuation. 5-13 - Ben Abramowitz, Omer Lev, Nicholas Mattei:
Who Reviews The Reviewers? A Multi-Level Jury Problem. 14-22 - Marek Adamczyk, Maurycy Borkowski, Michal Pawlowski:
An Improved Mechanism for Pricing Ride-Hailing Fares. 23-31 - Mahyar Afshinmehr, Alireza Danaei, Mehrafarin Kazemi, Kurt Mehlhorn, Nidhi Rathi:
EFX Allocations and Orientations on Bipartite Multi-graphs: A Complete Picture. 32-40 - Pulkit Agarwal, Harshvardhan Agarwal, Vaibhav Raj, Swaprava Nath:
Harmonious Balanced Partitioning of a Network of Agents. 41-49 - Rishabh Agrawal, Murtaza Asrani, Hadi Youssef, Apurva Narayan:
SCMRAG: Self-Corrective Multihop Retrieval Augmented Generation System for LLM Agents. 50-58 - Ohenewa Bediako Akuffo, Birgit Lugrin:
Investigating the Perspective of Non-Native Speakers on Foreigner-Directed Speech using Virtual Agents: The Role of Racial Ingroup Affiliation and Language Proficiency on Perception and Comprehension. 59-68 - Caren Al Anaissy, Jérôme Delobelle, Srdjan Vesic, Bruno Yun:
Impact Measures for Gradual Argumentation Semantics. 69-77 - Abu Mohammad Hammad Ali, Daniel Ogundare, Boting Yang, Sandra Zilles:
Approximation Ratio for Preference Aggregation Using Tree CP-Nets. 78-86 - Sharareh Alipour, Kajal Baghestani, Mahdis Mirzaei, Soroush Sahraei:
Geometric Freeze-Tag Problem. 87-95 - Shaull Almagor, Guy Avni, Neta Dafni:
Robin Hood Reachability Bidding Games. 96-104 - Abeer Alshehri, Hissah Alotaibi, Tim Miller, Mor Vered:
A Hypothesis-Driven Approach to Explainable Goal Recognition. 105-114 - Georgios Amanatidis, Georgios Birmpas, Philip Lazos, Stefano Leonardi, Rebecca Reiffenhäuser:
Algorithmically Fair Maximization of Multiple Submodular Objective Functions. 115-123 - Ramsundar Anandanarayanan, Swaprava Nath, Prasant Misra:
Truthful and Welfare-maximizing Resource Scheduling with Application to Electric Vehicles. 124-132 - Anastasia Sophia Apeiron, Davide Dell'Anna, Pradeep K. Murukannaiah, Pinar Yolum:
Model and Mechanisms of Consent for Responsible Autonomy. 133-141 - Leo Ardon, Daniel Furelos-Blanco, Roko Parac, Alessandra Russo:
FORM: Learning Expressive and Transferable First-Order Logic Reward Machines. 142-151 - Andreas Athanasopoulos, Anne-Marie George, Christos Dimitrakakis:
Probably Correct Optimal Stable Matching for Two-Sided Market Under Uncertainty. 152-160 - Guy Avni, Martin Kurecka, Kaushik Mallik, Petr Novotný, Suman Sadhukhan:
Bidding Games on Markov Decision Processes with Quantitative Reachability Objectives. 161-169 - Haris Aziz, Zixu He, Xinhang Lu, Kaiyang Zhou:
Fair Allocation of Divisible Goods under Non-Linear Valuations. 170-178 - Carmel Baharav, Andrei Constantinescu, Roger Wattenhofer:
Condorcet Winners and Anscombe's Paradox Under Weighted Binary Voting. 179-187 - Xiaoxin Bai, JunYang Yang, Shengchao Yuan, Yinghao Zhang, Hanqian Wu:
Local Topological Information as a Powerful Enhancer for Generalizable Neural Method in Travelling Salesman Problem. 188-196 - James P. Bailey, Craig A. Tovey:
On the Gale-Shapley Algorithm for Stable Matchings with a Partial Honesty Nash Refinement. 197-204 - James P. Bailey, Craig A. Tovey:
The Price of Anarchy in Spatial Social Choice. 205-213 - Gabriel Ballot, Vadim Malvone, Jean Leneutre, Jingxuan Ma, Mourad Leslous:
Alternating-time Temporal Logic with Stochastic Abilities. 214-222 - Michael Banck, Elisabeth Ganal, Hanna-Finja Weichert, Frank Puppe, Birgit Lugrin:
An AI-Driven Card Playing Robot: An Empirical Study on Communicative Style and Embodiment with Elderly Adults. 223-232 - Saptarashmi Bandyopadhyay, Mustafa Mert Çelikok, Robert Loftin:
On the Complexity of Learning to Cooperate in Populations of Socially Rational Agents. 233-241 - Tahsin Tariq Banna, Sejuti Rahman, Mohammad Tareq:
Beyond Words: Integrating Personality Traits and Context-Driven Gestures in Human-Robot Interactions. 242-251 - Kexin Gu Baugh, Luke Dickens, Alessandra Russo:
Neural DNF-MT: A Neuro-symbolic Approach for Learning Interpretable and Editable Policies. 252-260 - Divyagna Bavikadi, Nathaniel Lee, Paulo Shakarian, Chad Parvis:
Sea-cret Agents: Maritime Abduction for Region Generation to Expose Dark Vessel Trajectories. 261-270 - Petra Berenbrink, Martin Hoefer, Dominik Kaaser, Marten Maack, Malin Rau, Lisa Wilhelmi:
Opinion Dynamics with Median Aggregation. 271-279 - Sara Bernardini, Fabio Fagnani, Santiago Franco:
Speed vs Accuracy in Goal Recognition for Time-Sensitive Applications: A Game-Theoretic Approach. 280-288 - Damien Berriaud, Ezzat Elokda, Devansh Jalota, Emilio Frazzoli, Marco Pavone, Florian Dörfler:
To Spend or to Gain: Online Learning in Repeated Karma Auctions. 289-297 - Umang Bhaskar, Gunjan Kumar, Yeshwant Pandit, Rakshitha:
Towards Envy-Freeness Relaxations for General Nonmonotone Valuations. 298-306 - Umang Bhaskar, Juhi Chaudhary, Palash Dey:
Maximizing Value in Challenge the Champ Tournaments. 307-315 - Yu Bi, Lingxiao Zhao, Jinyun Tong, Zhe Feng, Carmine Ventre:
Agent-based Modeling and Simulation of Ambiguity in Catastrophe Insurance Markets. 316-324 - Martin Bichler, Markus Ewert, Axel Ockenfels:
Equilibrium Analysis in Markets with Asymmetric Utility Functions. 325-333 - Davide Bilò, Sarel Cohen, Tobias Friedrich, Hans Gawendowicz, Nicolas Klodt, Pascal Lenzner, George Skretas:
Temporal Network Creation Games: The Impact of Non-Locality and Terminals. 334-342 - Vittorio Bilò, Angelo Fanelli, Laurent Gourvès, Christos Tsoufis, Cosimo Vinci:
Minimizing Rosenthal's Potential in Monotone Congestion Games. 343-351 - Thomas Bläsius, Adrian Feilhauer, Markus Jung, Moritz Laupichler, Peter Sanders, Michael Zündorf:
Synergistic Traffic Assignment. 352-360 - Dominik Jacek Bogucki, Lukasz Lepak, Sonam Parashar, Bartlomiej Blachowski, Pawel Wawrzynski:
EnEnv 1.0: Energy Grid Environment for Multi-Agent Reinforcement Learning Benchmarking. 361-370 - Lorenzo Bonanni, Daniele Meli, Alberto Castellini, Alessandro Farinelli:
Monte Carlo Tree Search with Velocity Obstacles for Safe and Efficient Motion Planning in Dynamic Environments. 371-380 - Gusseppe Bravo Rocca, Peini Liu, Jordi Guitart, Rodrigo M. Carrillo-Larco, Ajay Dholakia, David Ellison:
Feature Engineering for Agents: An Adaptive Cognitive Architecture for Interpretable ML Monitoring. 381-389 - Robert Bredereck, Andrzej Kaczmarczyk, Junjie Luo, Bin Sun:
Computing Efficient Envy-Free Partial Allocations of Indivisible Goods. 390-398 - Asger Horn Brorholt, Kim Guldstrand Larsen, Christian Schilling:
Compositional Shielding and Reinforcement Learning for Multi-Agent Systems. 399-407 - Axel Brunnbauer, Julian Lemmel, Zahra Babaiee, Sophie A. Neubauer, Radu Grosu:
Scalable Offline Reinforcement Learning for Mean Field Games. 408-417 - Martin Bullinger, Vaggos Chatziafratis, Parnian Shahkar:
Welfare Approximation in Additively Separable Hedonic Games. 418-426 - Martin Bullinger, Edith Elkind, Mohamad Latifian:
Towards Fair and Efficient Public Transportation: A Bus Stop Model. 427-435 - Turgay Caglar, Sarath Sreedharan, Mor Vered:
Who Am I Dealing With? Explaining the Designer's Hidden Intentions. 436-444 - Katrin Casel, Stefan Neubert:
Emit As You Go: Enumerating Edges of a Spanning Tree. 445-453 - Karen Frilya Celine, Warut Suksompong, Sheung Man Yuen:
On the Fairness of Additive Welfarist Rules. 454-462 - T.-H. Hubert Chan, Qipeng Kuang, Quan Xue:
Game-Theoretically Secure Distributed Protocols for Fair Allocation in Coalitional Games. 463-471 - Harish Chandramouleeswaran, Prajakta Nimbhorkar, Nidhi Rathi:
Fair Division in a Variable Setting. 472-480 - Shenghui Chen, Ruihan Zhao, Sandeep Chinchali, Ufuk Topcu:
Human-Agent Coordination in Games under Incomplete Information via Multi-Step Intent. 481-489 - Amit K. Chopra, Matteo Baldoni, Samuel Christie, Munindar P. Singh:
Azorus: Commitments over Protocols for BDI Agents. 490-499 - Ayush Chopra, Shashank Kumar, Nurullah Giray Kuru, Ramesh Raskar, Arnau Quera-Bofarull:
On the Limits of Agency in Agent-based Models. 500-509 - Ari Conati, Andreas Niskanen, Ronald de Haan, Matti Järvisalo:
Computing Efficient and Envy-Free Allocations under Dichotomous Preferences using SAT. 510-518 - Andrei Constantinescu, Roger Wattenhofer:
Byzantine Game Theory: Sun Tzu's Boxes. 519-528 - Costas Courcoubetis, Antonis Dimakis:
Selfish Behavior and Resource Competition in Multi-Agent Systems. 529-537 - Gianlorenzo D'Angelo, Esmaeil Delfaraz:
Approximation Algorithms for Connected Maximum Coverage. 538-546 - Arpan Dasgupta, Gagan Jain, Arun Suggala, Karthikeyan Shanmugam, Milind Tambe, Aparna Taneja:
Bayesian Collaborative Bandits with Thompson Sampling for Improved Outreach in Maternal Health. 547-555 - Emanuele De Angelis, Maurizio Proietti, Francesca Toni:
Greedy ABA Learning for Case-Based Reasoning. 556-564 - Ali Safarpoor-Dehkordi, Ahad N. Zehmakan:
More Efficient Sybil Detection Mechanisms Leveraging Resistance of Users to Attack Requests. 565-573 - Florent Delgrange, Guy Avni, Anna Lukina, Christian Schilling, Ann Nowé, Guillermo A. Pérez:
Composing Reinforcement Learning Policies, with Formal Guarantees. 574-583 - Argyrios Deligkas, Eduard Eiben, Robert Ganian, Iyad Kanj, M. S. Ramanujan:
Parameterized Algorithms for Multiagent Pathfinding on Trees. 584-592 - Shilong Deng, Yongzhao Wang, Rahul Savani:
From Natural Language to Extensive-Form Game Representations. 593-601 - Anthony DiGiovanni, Jesse Clifton, Nicolas Macé:
Safe Pareto Improvements for Expected Utility Maximizers in Program Games. 602-610 - Jonathan Diller, Qi Han, Robert Byers, James Dotterweich, James Humann:
Hitchhiker's Guide to Patrolling: Path-Finding for Energy-Sharing Drone-UGV Teams. 611-619 - Youcef Djenouri, Nassim Belmecheri, Tomasz P. Michalak, Jan Dubinski, Ahmed Nabil Belbachir, Anis Yazidi:
Learning Graph Representation of Agent Diffusers. 620-629 - Chris Dong, Martin Bullinger, Tomasz Was, Larry Birnbaum, Edith Elkind:
Selecting Interlacing Committees. 630-638 - Wen Dong, Fairul Mohd-Zaid:
Simulating and Evaluating Generative Modeling and Collaborative Filtering in Complex Social Networks. 639-648 - Yuriy Dorn, Aleksandr Katrutsa, Ilgam Latypov, Andrey Pudovikov:
Fast UCB-type Algorithms for Stochastic Bandits with Heavy and Super Heavy Symmetric Noise. 649-657 - François Durand:
Why Instant-Runoff Voting Is So Resilient to Coalitional Manipulation: Phase Transitions in the Perturbed Culture. 658-666 - Soroush Ebadian, Evi Micha:
Boosting Sortition via Proportional Representation. 667-675 - Edith Elkind, Alexander Lam, Mohamad Latifian, Tzeh Yuan Neoh, Nicholas Teh:
Temporal Fair Division of Indivisible Items. 676-685 - Thorsten Engesser, Thibaut Le Marre, Emiliano Lorini, François Schwarzentruber, Bruno Zanuttini:
A Simple Integration of Epistemic Logic and Reinforcement Learning. 686-694 - Emre Erdogan, Hüseyin Aydin, Frank Dignum, Rineke Verbrugge, Pinar Yolum:
Mitigating Value Conflicts with Computational Theory of Mind. 695-703 - Piotr Faliszewski, Lukasz Janeczko, Andrzej Kaczmarczyk, Marcin Kurdziel, Grzegorz Pierczynski, Stanislaw Szufa:
Learning Real-Life Approval Elections. 704-712 - Flint Xiaofeng Fan, Cheston Tan, Yew-Soon Ong, Roger Wattenhofer, Wei Tsang Ooi:
FedRLHF: A Convergence-Guaranteed Federated Framework for Privacy-Preserving and Personalized RLHF. 713-721 - Liangda Fang, Shikang Chen, Xiaoman Wang, Xiaoyou Lin, Chenyi Zhang, Qingliang Chen, Quanlong Guan, Kaile Su:
Automatic Verification of Linear Integer Planning Programs via Forgetting in LIAUPF. 722-730 - Xing Fang, Qichao Zhang, Haoran Li, Dongbin Zhao:
Consistency Policy with Categorical Critic for Autonomous Driving. 731-739 - Qihui Feng, Hannah Wilk, Shakil M. Khan, Gerhard Lakemeyer:
Translating Multi-Agent Modal Logics of Knowledge and Belief into Decidable First-Order Fragments. 740-748 - Foivos Fioravantes, Abhiruk Lahiri, Antonio Lauerbach, Lluís Sabater, Marie Diana Sieper, Samuel Wolf:
Eliminating Majority Illusion. 749-757 - Zack Fitzsimmons, Vignesh Viswanathan, Yair Zick:
On the Hardness of Fair Allocation under Ternary Valuations. 758-766 - Michele Flammini, Maria Fomenko, Giovanna Varricchio:
Non-obvious Manipulability in Hedonic Games with Friends Appreciation Preferences. 767-775 - Jack Foxabbott, Rohan Subramani, Francis Rhys Ward:
Higher-Order Belief in Incomplete Information MAIDs. 776-784 - Fabian Frank, Patrick Lederer:
The Metric Distortion of Randomized Social Choice Functions: C1 Maximal Lottery Rules and Simulations. 785-793 - Rupert Freeman, Geoffrey Pritchard, Mark C. Wilson:
Order Symmetry: A New Fairness Criterion for Assignment Mechanisms. 794-802 - Paul Friedrich, Barna Pásztor, Giorgia Ramponi:
Learning Collusion in Episodic, Inventory-Constrained Markets. 803-812 - Yuma Fujimoto, Kaito Ariu, Kenshi Abe:
Global Behavior of Learning Dynamics in Zero-Sum Games with Memory Asymmetry. 813-819 - Pranshu Gaba, Shibashis Guha:
Optimising Expectation with Guarantees for Window Mean Payoff in Markov Decision Processes. 820-828 - Rustam Galimullin, Maksim Gladyshev, Munyque Mittelmann, Nima Motamed:
Changing the Rules of the Game: Reasoning About Dynamic Phenomena in Multi-Agent Systems. 829-838 - Rohan Garg, Alexandros Psomas:
Fairly Allocating Goods in Parallel. 839-847 - Marcelo Matheus Gauy, Anna Abramishvili, Eduardo Colli, Tiago Madeira, Frederik Mallmann-Trenn, Vinícius Franco Vasconcelos, David Kohan Marzagão:
Voter Model Meets Rumour Spreading: A Study of Consensus Protocols on Graphs with Agnostic Nodes. 848-857 - Zichang Ge, Changyu Chen, Arunesh Sinha, Pradeep Varakantham:
On Learning Informative Trajectory Embeddings for Imitation, Classification and Regression. 858-866 - Minghong Geng, Shubham Pateria, Budhitama Subagdja, Ah-Hwee Tan:
MOSMAC: A Multi-agent Reinforcement Learning Benchmark on Sequential Multi-Objective Tasks. 867-876 - Francesco Giacomarra, Mehran Hosseini, Nicola Paoletti, Francesca Cairoli:
Certified Guidance for Planning with Deep Generative Models. 877-885 - Roman Chiva Gil, Daniel Jarne Ornia, Khaled A. Mustafa, Javier Alonso-Mora:
Predictability Awareness for Efficient and Robust Multi-Agent Coordination. 886-894 - Hugo Gimbert, Soumyajit Paul, B. Srivathsan:
Simplifying Imperfect Recall Games. 895-903 - Victor Gimenez-Abalos, Sergio Álvarez-Napagao, Adrián Tormos, Ulises Cortés, Javier Vázquez-Salceda:
Policy Graphs and Intention: Answering 'Why' and 'How' from a Telic Perspective. 904-913 - Salil Gokhale, Harshul Sagar, Rohit Vaish, Jatin Yadav:
Approximating One-Sided and Two-Sided Nash Social Welfare With Capacities. 914-922 - Sreenivas Gollapudi, Kostas Kollias, Alkmini Sgouritsa, Ali Kemal Sinop:
Fairness and Optimality in Routing. 923-931 - Victor Guillet, Charles Lesire, Gauthier Picard, Christophe Grand:
Extending Consensus-based Task Allocation Algorithms with Bid Intercession to Foster Mixed-Initiative. 932-940 - Swapnil Guragain, Gokarna Sharma:
On the Power of Temporal Locality on Online Routing Problems. 941-949 - Sabit Hassan, Hye-Young Chung, Xiang Zhi Tan, Malihe Alikhani:
Coherence-Driven Multimodal Safety Dialogue with Active Learning for Embodied Agents. 950-959 - Somnath Hazra, Pallab Dasgupta, Soumyajit Dey:
Tackling Uncertainties in Multi-Agent Reinforcement Learning through Integration of Agent Termination Dynamics. 960-968 - Benjamin Heymann, Marc Lanctot:
Learning in Games with Progressive Hiding. 969-977 - Mehran Hosseini, Alessio Lomuscio, Nicola Paoletti:
LTL Verification of Memoryful Neural Agents. 978-987 - Vincent Hsiao, Mark Roberts, Laura M. Hiatt, George Dimitri Konidaris, Dana S. Nau:
Automating Curriculum Learning for Reinforcement Learning using a Skill-Based Bayesian Network. 988-996 - Kun Hu, Muning Wen, Xihuai Wang, Shao Zhang, Yiwei Shi, Minne Li, Minglong Li, Ying Wen:
PMAT: Optimizing Action Generation Order in Multi-Agent Reinforcement Learning. 997-1005 - Yiting Hu, Lingjie Duan:
Truthful Mechanisms for Linear Bandit Games with Private Contexts. 1006-1014 - Chuanbo Hua, Federico Berto, Jiwoo Son, Seunghyun Kang, Changhyun Kwon, Jinkyoo Park:
CAMP: Collaborative Attention Model with Profiles for Vehicle Routing Problems. 1015-1024 - Maxence Hussonnois, Thommen George Karimpanal, Santu Rana:
Human-Aligned Skill Discovery: Balancing Behaviour Exploration and Alignment. 1025-1033 - Lihi Idan, Ming Li:
Responsible Uplift Modeling. 1034-1041 - Taher Jafferjee, Juliusz Ziomek, Tianpei Yang, Zipeng Dai, Jianhong Wang, Matthew E. Taylor, Kun Shao, Jun Wang, David Mguni:
Taming Multi-Agent Reinforcement Learning with Estimator Variance Reduction. 1042-1050 - Wojciech Jamroga, Marta Kwiatkowska, Wojciech Penczek, Laure Petrucci, Teofil Sidoruk:
Probabilistic Timed ATL. 1051-1059 - Jiaxuan Jiang, Ling Pan, Lin Zhou, Longbo Huang, Zhixuan Fang:
Tackling Sparsity in Designated Driver Dispatch with Multi-Agent Reinforcement Learning. 1060-1069 - Yusuf Hakan Kalayci, Jiasen Liu, David Kempe:
Full Proportional Justified Representation. 1070-1078 - Shivaram Kalyanakrishnan, Sheel Shah, Santhosh Kumar Guguloth:
A View of the Certainty-Equivalence Method for PAC RL as an Application of the Trajectory Tree Method. 1079-1087 - Benjamin Kempinski, Ian Gemp, Kate Larson, Marc Lanctot, Yoram Bachrach, Tal Kachman:
Game of Thoughts: Iterative Reasoning in Game-Theoretic Domains with Large Language Models. 1088-1097 - Sylvia S. Kerkhove, Natasha Alechina, Mehdi Dastani:
Causes and Strategies in Multiagent Systems. 1098-1106 - Sungsik Kim, Janghyun Baek, Jinkyu Kim, Jaekoo Lee:
GUIDE-CoT: Goal-driven and User-Informed Dynamic Estimation for Pedestrian Trajectory using Chain-of-Thought. 1107-1116 - Yan Kim, Wojciech Jamroga, Wojciech Penczek, Laure Petrucci:
Practical Abstractions for Model Checking Continuous-Time Multi-Agent Systems. 1117-1126 - Fatih Erdem Kizilkaya, David Kempe:
k-ApprovalVeto: A Spectrum of Voting Rules Balancing Metric Distortion and Minority Protection. 1127-1135 - Yoshikatsu Kobayashi, Koji Hasebe:
Robustness of Epistemic Gossip Protocols Against Data Loss. 1136-1144 - Natalia Koliou, George A. Vouros:
Ranking Joint Policies in Dynamic Games using Evolutionary Dynamics. 1145-1153 - Simon Kolker, Louise A. Dennis, Ramon Fraga Pereira, Mengwei Xu:
Uncertain Machine Ethics Planning. 1154-1162 - Christine Konicki, Mithun Chakraborty, Michael P. Wellman:
Policy Abstraction and Nash Refinement in Tree-Exploiting PSRO. 1163-1171 - Avinash Kori, Antonio Rago, Francesca Toni:
Free Argumentative Exchanges for Explaining Image Classifiers. 1172-1180 - Qian Kou, Mingyang Li, Zeyang Liu, Long Qian, Zhuoran Chen, Lipeng Wan, Xingyu Chen, Xuguang Lan:
Offline Multi-Agent Preference-based Reinforcement Learning with Agent-aware Direct Preference Optimization. 1181-1190 - Vojtech Kovarík, Nathaniel Sauerberg, Lewis Hammond, Vincent Conitzer:
Game Theory with Simulation in the Presence of Unpredictable Randomisation. 1191-1199 - Merlijn Krale, Wietze Koops, Sebastian Junges, Thiago D. Simão, Nils Jansen:
Tighter Value-Function Approximations for POMDPs. 1200-1208 - Simon Krogmann, Pascal Lenzner, Alexander Skopalik:
The Bakers and Millers Game with Restricted Locations. 1209-1217 - Ajay D. Kshemkalyani, Manish Kumar, Anisur Rahaman Molla, Gokarna Sharma:
Near-Linear Time Leader Election in Multiagent Networks. 1218-1226 - Abhishek Ninad Kulkarni, Andy Liu, Jean-Raphaël Gaglione, Daniel Fried, Ufuk Topcu:
Dynamic Coalition Structure Detection in Natural-Language-based Interactions. 1227-1234 - Vikas Kumar, Ajin George Joseph:
Emergence of Recursive Language through Bootstrapping and Iterated Learning. 1235-1243 - Hemant Kumawat, Saibal Mukhopadhyay:
AdaCred: Adaptive Causal Decision Transformers with Feature Crediting. 1244-1252 - Marc Lanctot, Kate Larson, Michael Kaisers, Quentin Berthet, Ian Gemp, Manfred Diaz, Roberto-Rafael Maura-Rivero, Yoram Bachrach, Anna Koop, Doina Precup:
Soft Condorcet Optimization for Ranking of General Agents. 1253-1262 - Sunbowen Lee, Hongqin Lyu, Yicheng Gong, Yingying Sun, Chao Deng:
MacLight: Multi-scene Aggregation Convolutional Learning for Traffic Signal Control. 1263-1271 - Jean Leneutre, Vadim Malvone, James Ortiz:
Timed Obstruction Logic: A Timed Approach to Dynamic Game Reasoning. 1272-1281 - Chin-wing Leung, Paolo Turrini, Ann Nowé:
Curiosity-Driven Partner Selection Accelerates Convention Emergence in Language Games. 1282-1290 - Tianxu Li, Kun Zhu:
Self-Supervised Multi-Agent Diversity with Nonparametric Entropy Maximization. 1291-1299 - Xiaohan Li, Wenxiang Shen, Dong Liu, Jun Wu:
OGS-SLAM: Hybrid ORB-Gaussian Splatting SLAM. 1300-1308 - Yinfeng Li, Emiliano Lorini, Munyque Mittelmann:
Rational Capability in Concurrent Games. 1309-1317 - Yugu Li, Zehong Cao, Jianglin Qiao, Siyi Hu:
Nucleolus Credit Assignment for Effective Coalitions in Multi-agent Reinforcement Learning. 1318-1326 - Wei-Chen Liao, Ti-Rong Wu, I-Chen Wu:
Dynamic Sight Range Selection in Multi-Agent Reinforcement Learning. 1327-1335 - Xiaoshan Lin, Roberto Tron:
Adaptive Bi-Level Multi-Robot Task Allocation and Learning under Uncertainty with Temporal Logic Constraints. 1336-1344 - Fangqi Liu, Rishav Sen, Jose Paolo Talusan, Ava Pettet, Aaron Kandel, Yoshinori Suzue, Ayan Mukhopadhyay, Abhishek Dubey:
Reinforcement Learning-based Approach for Vehicle-to-Building Charging with Heterogeneous Agents and Long Term Rewards. 1345-1353 - Jiayuan Liu, Siwei Wang, Zhixuan Fang:
Efficient and Optimal Policy Gradient Algorithm for Corrupted Multi-armed Bandits. 1354-1361 - Siyu Liu, Rida A. Bazzi, Fei Fang, Tiffany Bao:
Teamwork Makes the Defense Work: Comprehensive Vulnerability Defense Resource Allocation. 1362-1370 - Siyuan Liu, Yulong Zeng:
Games in Public Announcement: How to Reduce System Losses in Optimistic Blockchain Mechanisms. 1371-1379 - Yiping Liu, Mengxiao Zhang, Jiamou Liu, Song Yang:
Data Pricing for Graph Neural Networks without Pre-purchased Inspection. 1380-1388 - Zeyuan Liu, Zhirui Fang, Jiafei Lyu, Xiu Li:
Leveraging Score-based Models for Generating Penalization in Model-based Offline Reinforcement Learning. 1389-1398 - Donald Loveland, James Usevitch, Zachary Serlin, Danai Koutra, Rajmonda Caceres:
MAGNET: A Multi-Agent Graph Neural Network for Efficient Bipartite Task Assignment. 1399-1407 - Kun Ma, Qilong Han, Jingzheng Yao:
Multi-Ship Future Interaction Trajectory Prediction via Pre-Initializer Diffusion Model. 1408-1417 - Amir Maliah, Dor Atzmon, Ariel Felner:
Minimizing Makespan with Conflict-Based Search for Optimal Multi-Agent Path Finding. 1418-1426 - Sheryl Mantik, Michael Dann, Minyi Li, Huong Ha, Julie Porteous:
Beyond Goal Recognition: A Reinforcement Learning-based Approach to Inferring Agent Behaviour. 1427-1435 - Piyushi Manupriya, Himanshu, Saketha Nath Jagarlapudi, Ganesh Ghalme:
Multi-agent Multi-armed Bandits with Minimum Reward Guarantee Fairness. 1436-1444 - Enrico Marchesini, Andrea Baisero, Rupali Bhati, Christopher Amato:
On Stateful Value Factorization in Multi-Agent Reinforcement Learning. 1445-1453 - Carlos Martin, Tuomas Sandholm:
ApproxED: Approximate Exploitability Descent via Learned Best Responses. 1454-1463 - Luca Marzari, Priya L. Donti, Changliu Liu, Enrico Marchesini:
Improving Policy Optimization via ε-Retrain. 1464-1472 - Connor Mattson, Varun Raveendra, Ricardo Vega, Cameron Nowzari, Daniel S. Drew, Daniel S. Brown:
Discovery and Deployment of Emergent Robot Swarm Behaviors via Representation Learning and Real2Sim2Real Transfer. 1473-1482 - Felipe Meneguzzi, Ramon Fraga Pereira, Nir Oren:
Generalised BDI Planning. 1483-1491 - Benedikt Valentin Meylahn:
Multi-agent Reinforcement Learning in the All-or-Nothing Public Goods game on Networks. 1492-1500 - Kartik Nagpal, Dayi Dong, Negar Mehr:
Leveraging Large Language Models for Effective and Explainable Multi-Agent Credit Assignment. 1501-1510 - Sanjeev Nahulanthran, Leimin Tian, Dana Kulic, Mor Vered:
Explaining Facial Expression Recognition. 1511-1519 - Natinael Solomon Neggatu, Jeremie Houssineau, Giovanni Montana:
Evaluation-Time Policy Switching for Offline Reinforcement Learning. 1520-1528 - Jessica L. Newman, Enrico H. Gerding, Enrico Marchioni, Baharak Rastegari:
Resource Task Games. 1529-1537 - Lewis Newsham, Daniel Prince:
Personality-Driven Decision Making in LLM-Based Autonomous Agents. 1538-1547 - Dung Nguyen, Ariel Vetzler, Sarit Kraus, Anil Vullikanti:
Contrastive Explainable Clustering with Differential Privacy. 1548-1556 - Kieu Thao Nguyen Pham, Rachael Hwee Ling Sim, Quoc Phong Nguyen, See-Kiong Ng, Bryan Kian Hsiang Low:
DUPRE: Data Utility Prediction for Efficient Data Valuation. 1557-1565 - Erfaun Noorani, Pasan Dissanayake, Faisal Hamman, Sanghamitra Dutta:
Counterfactual Explanations for Model Ensembles Using Entropic Risk Measures. 1566-1575 - Helbert Paat, Guohao Shen:
Conformal Set-based Human-AI Complementarity with Multiple Experts. 1576-1585 - Aritra Pal, Anandsingh Chauhan, Mayank Baranwal:
Together We Rise: Optimizing Real-Time Multi-Robot Task Allocation using Coordinated Heterogeneous Plays. 1586-1594 - Fernando Palafox, Jesse Milzman, Dong Ho Lee, Ryan Park, David Fridovich-Keil:
Smooth Information Gathering in Two-Player Noncooperative Games. 1595-1603 - Yuxin Pan, Ruohong Liu, Yize Chen, Zhiguang Cao, Fangzhen Lin:
Hierarchical Learning-based Graph Partition for Large-scale Vehicle Routing Problems. 1604-1612 - George Papadopoulos, Andreas Kontogiannis, Foteini Papadopoulou, Chaido Poulianou, Ioannis Koumentis, George A. Vouros:
An Extended Benchmarking of Multi-Agent Reinforcement Learning Algorithms in Complex Fully Cooperative Tasks. 1613-1622 - Kexing Peng, Pengyi Li, Jianye Hao:
Enhancing Graph-based Coordination with Evolutionary Algorithms for Episodic Multi-agent Reinforcement Learning. 1623-1631 - Nianli Peng, Muhang Tian, Brandon Fain:
Multi-objective Reinforcement Learning with Nonlinear Preferences: Provable Approximation for Maximizing Expected Scalarized Return. 1632-1640 - Quang Anh Pham, Janaka Chathuranga Brahmanage, Akshat Kumar:
ShipNaviSim: Data-Driven Simulation for Real-World Maritime Navigation. 1641-1649 - Alexandre S. Pires, Fernando P. Santos:
Artificial Agents Mitigate The Punishment Dilemma Of Indirect Reciprocity. 1650-1659 - Subham Pokhriyal, Shweta Jain, Ganesh Ghalme, Vaneet Aggarwal:
Anytime Fairness Guarantees in Stochastic Combinatorial MABs: A Novel Learning Framework. 1660-1669 - Antigoni Polychroniadou, T.-H. Hubert Chan, Adya Agrawal:
Indifferential Privacy: A New Paradigm and Its Applications to Optimal Matching in Dark Pool Auctions. 1670-1678 - Koen Ponse, Aske Plaat, Niki van Stein, Thomas M. Moerland:
EconoJax: A Fast & Scalable Economic Simulation in JAX. 1679-1687 - Francesco Pontiggia, Filip Macák, Roman Andriushchenko, Michele Chiari, Milan Ceska:
Decentralized Planning Using Probabilistic Hyperproperties. 1688-1697 - David Porfirio, Mark Roberts, Laura M. Hiatt:
Uncertainty Expression for Human-Robot Task Communication. 1698-1707 - Nikhilesh Prabhakar, Ranveer Singh, Harsha Kokel, Sriraam Natarajan, Prasad Tadepalli:
Combining Planning and Reinforcement Learning for Solving Relational Multiagent Domains. 1708-1717 - Nathan Qiu, Daniel Liang:
Reinforcement Learning Based Simulated Annealing. 1718-1726 - Gregg R. Rabideau, Joseph A. Russino, Andrew Branch, Nihal Dhamani, Tiago Stegun Vaquero, Steve A. Chien, Jean-Pierre de la Croix, Federico Rossi:
Planning, Scheduling, and Execution on the Moon: The CADRE Technology Demonstration Mission. 1727-1735 - Hassan Raissouni, Wissal Bekhti, Btissam El Khamlichi, Amal El Fallah Seghrouchni:
Reputation-Filtered Reward Reshaping: Encouraging Cooperation in High Dimensional Semi-Cooperative Multi-agent Settings. 1736-1744 - Tianyu Ren, Xuan Yao, Yang Li, Xiao-Jun Zeng:
Bottom-Up Reputation Promotes Cooperation with Multi-Agent Reinforcement Learning. 1745-1754 - Jennifer Renoux, Filipa Correia, Joana Campos, Lucas Morillo-Mendez, Neziha Akalin, Fernando P. Santos, Ana Paiva:
The Effect of Agent-based Feedback on Prosociality in Social Dilemmas. 1755-1763 - Anna Riedmann, Carlo D'Eramo, Birgit Lugrin:
Real-World Testing Matters in Reinforcement Learning for Education. 1764-1773 - Willem Röpke, Mathieu Reymond, Patrick Mannion, Diederik M. Roijers, Ann Nowé, Roxana Radulescu:
Divide and Conquer: Provably Unveiling the Pareto Front with Multi-Objective Reinforcement Learning. 1774-1783 - Daniel J. Rosenkrantz, Madhav V. Marathe, Zirou Qiu, S. S. Ravi, Richard Edwin Stearns:
On Some Fundamental Problems for Multi-Agent Systems Over Multilayer Networks. 1784-1792 - Jaime Ruiz-Serra, Patrick Sweeney, Michael S. Harré:
Factorised Active Inference for Strategic Multi-Agent Interactions. 1793-1802 - Pulkit Rustagi, Yashwanthi Anand, Sandhya Saisubramanian:
Multi-Objective Planning with Contextual Lexicographic Reward Preferences. 1803-1811 - Fardin Saad, Pradeep K. Murukannaiah, Munindar P. Singh:
Gricean Norms as a Basis for Effective Collaboration. 1812-1820 - Huma Samin, Dylan J. Walton, Nelly Bencomo:
Surprise! Surprise! Learn and Adapt. 1821-1829 - Bidipta Sarkar, Warren Xia, C. Karen Liu, Dorsa Sadigh:
Training Language Models for Social Deduction with Multi-Agent Reinforcement Learning. 1830-1839 - Camilo Sarmiento, Gauvain Bourgne, Jean-Gabriel Ganascia:
Formalising Overdetermination in a Labelled Transition System. 1840-1848 - Lukas Schäfer, Oliver Slumbers, Stephen McAleer, Yali Du, Stefano V. Albrecht, David Mguni:
Ensemble Value Functions for Efficient Exploration in Multi-Agent Reinforcement Learning. 1849-1857 - Ildikó Schlotter, Katarína Cechlárová:
Candidate Nomination for Condorcet-consistent Voting Rules. 1858-1866 - Ildikó Schlotter, Lydia Mirabel Mendoza-Cadena:
The Strong Core of Housing Markets with Partial Order Preferences. 1867-1875 - Sangwon Seo, Bing Han, Rayan Ebnali Harari, Roger D. Dias, Marco A. Zenati, Eduardo Salas, Vaibhav V. Unhelkar:
Socratic: Enhancing Human Teamwork via AI-enabled Coaching. 1876-1885 - Sangwon Seo, Vaibhav V. Unhelkar:
Hierarchical Imitation Learning of Team Behavior from Heterogeneous Demonstrations. 1886-1894 - Lorenzo Serina, Mattia Chiari, Alfonso Emilio Gerevini, Luca Putelli, Ivan Serina:
Towards Efficient Online Goal Recognition through Deep Learning. 1895-1903 - Ameesh Shah, Niklas Lauffer, Thomas Chen, Nikhil Pitta, Sanjit A. Seshia:
Learning Symbolic Task Decompositions for Multi-Agent Teams. 1904-1913 - Junning Shao, Siwei Wang, Zhixuan Fang:
Learning with Limited Shared Information in Multi-agent Multi-armed Bandit. 1914-1922 - Yuming Shao, Zhixuan Fang:
Incentivizing Truth Exploration and Honest Reporting: A Contract Design Approach. 1923-1931 - Risal Shahriar Shefin, Md Asifur Rahman, Thai Le, Sarra M. Alqahtani:
xSRL: Safety-Aware Explainable Reinforcement Learning - Safety as a Product of Explainability. 1932-1940 - David Shoresh, Yonatan Loewenstein:
Modeling the Centaur: Human-Machine Synergy in Sequential Decision Making. 1941-1949 - Davide Soldà, Pedro Cabalar, Agata Ciabattoni, Emery A. Neufeld:
Tackling Temporal Deontic Challenges with Equilibrium Logic. 1950-1958 - Xinwei Song, Tianyi Yang, Dengji Zhao:
Housing Market on Networks. 1959-1967 - Julien Soulé, Jean-Paul Jamont, Michel Occello, Louis-Marie Traonouez, Paul Théron:
An Organizationally-Oriented Approach to Enhancing Explainability and Control in Multi-Agent Reinforcement Learning. 1968-1976 - Chuxiong Sun, Peng He, Rui Wang, Changwen Zheng:
Revisiting Communication Efficiency in Multi-Agent Reinforcement Learning from the Dimensional Analysis Perspective. 1977-1986 - Jingbo Sun, Songjun Tu, Qichao Zhang, Ke Chen, Dongbin Zhao:
Salience-Invariant Consistent Policy Learning for Generalization in Visual Reinforcement Learning. 1987-1995 - Maciej Swiechowski, Dominik Slezak:
The Many Challenges of Human-Like Agents in Virtual Game Environments. 1996-2005 - Sheikh A. Tahmid, Gennaro Notomista:
Value Iteration for Learning Concurrently Executable Robotic Control Tasks. 2006-2014 - Mohamed S. Talamali, Genki Miyauchi, Thomas Watteyne, Micael S. Couceiro, Roderich Groß:
Ready, Bid, Go! On-Demand Delivery Using Fleets of Drones with Unknown, Heterogeneous Energy Storage Constraints. 2015-2023 - Biaoshuai Tao, Chengkai Zhang, Houyu Zhou:
The Degree of (Extended) Justified Representation and Its Optimization. 2024-2032 - Jia Tao, Xinran Zhang:
Logic of Knowledge and Cognitive Ability. 2033-2041 - Sidney Tio, Dexun Li, Pradeep Varakantham:
EduQate: Generating Adaptive Curricula through RMABs in Education Settings. 2042-2050 - Parisa Ghanad Torshizi, Laura B. Hensel, Ari Shapiro, Stacy C. Marsella:
Large Language Models for Virtual Human Gesture Selection. 2051-2059 - Dimitrios Troullinos, Georgios Chalkiadakis, Ioannis Papamichail, Markos Papageorgiou:
Conditional Max-Sum for Asynchronous Multiagent Decision Making. 2060-2068 - Songjun Tu, Jingbo Sun, Qichao Zhang, Xiangyuan Lan, Dongbin Zhao:
Online Preference-based Reinforcement Learning with Self-augmented Feedback from Large Language Model. 2069-2077 - Filip Úradník, Amanda Wang, Jie Gao:
Maximizing Truth Learning in a Social Network is NP-hard. 2078-2086 - Guilherme S. Varela, Alberto Sardinha, Francisco S. Melo:
Networked Agents in the Dark: Team Value Learning under Partial Observability. 2087-2095 - Kryspin Varys, Federico Cerutti, Adam J. Sobey, Timothy J. Norman:
HAVA: Hybrid Approach to Value-Alignment through Reward Weighing for Reinforcement Learning. 2096-2104 - Victor Villin, Thomas Kleine Buening, Christos Dimitrakakis:
A Minimax Approach to Ad Hoc Teamwork. 2105-2114 - Fábio Vital, Alberto Sardinha, Francisco S. Melo:
Implicit Repair with Reinforcement Learning in Emergent Communication. 2115-2124 - Avyukta Manjunatha Vummintala, Shivam Gupta, Shweta Jain, Sujit Gujar:
FLIGHT: Facility Location Integrating Generalized, Holistic Theory of Welfare. 2125-2133 - Peng-Yuan Wang, Jing-Cheng Pang, Chenyang Wang, Xu-Hui Liu, Tian-Shuo Liu, Si-Hang Yang, Hong Qian, Yang Yu:
InCLET: Large Language Model In-context Learning can Improve Embodied Instruction-following. 2134-2142 - Tonghan Wang, Heng Dong, Yanchen Jiang, David C. Parkes, Milind Tambe:
On Diffusion Models for Multi-Agent Partial Observability: Shared Attractors, Error Bounds, and Composite Flow. 2143-2152 - Xinghai Wei, Tingting Yuan, Jie Yuan, Dongxiao Liu, Xiaoming Fu:
ReSCOM: Reward-Shaped Curriculum for Efficient Multi-Agent Communication Learning. 2153-2161 - Jiaqi Wen, Leonardo Amado:
Goal Recognition via Variational Causality. 2162-2170 - Michael Winikoff, John Thangarajah, Sebastian Rodriguez:
A Scoresheet for Explainable AI. 2171-2180 - Hang Xiao, Huale Li, Shuhan Qi, Jiajia Zhang, Dingzhong Cai:
FGLight: Learning Neighbor-level Information for Traffic Signal Control. 2181-2189 - Yi Xie, Ziqing Zhou, Chun Ouyang, Siao Liu, Linqiang Hu, Zhongxue Gan:
ACORN: Acyclic Coordination with Reachability Network to Reduce Communication Redundancy in Multi-Agent Systems. 2190-2198 - Guojun Xiong, Haichuan Wang, Yuqi Pan, Saptarshi Mandal, Sanket Shah, Niclas Boehmer, Milind Tambe:
Finite-Horizon Single-Pull Restless Bandits: An Efficient Index Policy For Scarce Resource Allocation. 2199-2207 - Yiqing Xu, Finale Doshi-Velez, David Hsu:
On the Effective Horizon of Inverse Reinforcement Learning. 2208-2216 - Likun Yang, Pei Xu, Shiyue Cao, Yongjian Ren, Xiaotang Chen, Kaiqi Huang:
Uncertainty-Aware Opponent Modeling for Deep Reinforcement Learning. 2217-2225 - Yaodong Yang, Guangyong Chen, Hongyao Tang, Furui Liu, Danruo Deng, Pheng-Ann Heng:
Dual Ensembled Multiagent Q-Learning with Hypernet Regularizer. 2226-2234 - Yue Yang, Fan Yang, Yu Bai, Hao Wang:
Self-Interpretable Reinforcement Learning via Rule Ensembles. 2235-2243 - Tomohiko Yokoyama, Ayumi Igarashi:
Asymptotic Existence of Class Envy-free Matchings. 2244-2252 - Byunghyun Yoo, Younghwan Shin, Hyunwoo Kim, Euisok Chung, Jeongmin Yang:
Adaptive Episode Length Adjustment for Multi-agent Reinforcement Learning. 2253-2261 - Peihong Yu, Manav Mishra, Syed Zaidi, Pratap Tokekar:
Task-Agnostic Contrastive pre-Training for Inter-Agent Communication. 2262-2270 - Xingrui Yu, Zhenglin Wan, David Mark Bossens, Yueming Lyu, Qing Guo, Ivor W. Tsang:
Imitation from Diverse Behaviors: Wasserstein Quality Diversity Imitation Learning with Single-Step Archive Exploration. 2271-2280 - Uriel Zaed, Omer Lev, Roie Zivan:
Insights Regarding the Success of Damping in Improving Belief Propagation. 2281-2289 - Sheng Zang, Zhiguang Cao, Bo An, Senthilnath Jayavelu, Xiaoli Li:
Enhancing Sub-Optimal Trajectory Stitching: Spatial Composition RvS for Offline RL. 2290-2298 - Zehua Zang, Chuxiong Sun, Lixiang Liu, Fuchun Sun, Changwen Zheng:
Loss of Plasticity: A New Perspective on Solving Multi-Agent Exploration for Sparse Reward Tasks. 2299-2308 - Jinghan A. Zeng, Ruta Mehta:
On the Structure of EFX Orientations on Graphs. 2309-2316 - Hongming Zhang, Fengshuo Bai, Chenjun Xiao, Chao Gao, Bo Xu, Martin Müller:
β-DQN: Improving Deep Q-Learning By Evolving the Behavior. 2317-2326 - Junyu Zhang, Yao Zhang, Yaoxin Ge, Dengji Zhao, Hu Fu, Zhihao Gavin Tang, Pinyan Lu:
Incentives for Early Arrival in Cost Sharing. 2327-2335 - Yaocheng Zhang, Yuanheng Zhu, Yuqian Fu, Songjun Tu, Dongbin Zhao:
Offline Goal-Conditioned Reinforcement Learning with Elastic-Subgoal Diffused Policy Learning. 2336-2344 - Zeren Zhang, Zhiwei Xu, Guangchong Zhou, Dapeng Li, Bin Zhang, Guoliang Fan:
Unveiling Decision Intention for Cooperative Multi-Agent Reinforcement Learning. 2345-2354 - Lingxiao Zhao, Maria Polukarov, Carmine Ventre:
Agent-Based Analysis of Green Disclosure Policies and Their Market-Wide Impact on Firm Behavior. 2355-2363 - Zhiyu Zhao, Chengdong Ma, Qirui Mi, Ning Yang, Xue Yan, Mengyue Yang, Haifeng Zhang, Jun Wang, Yaodong Yang:
Mean Field Correlated Imitation Learning. 2364-2372 - Hai Zhong, Xun Wang, Zhuoran Li, Longbo Huang:
Offline-to-Online Multi-Agent Reinforcement Learning with Offline Value Function Memory and Sequential Exploration. 2373-2381 - Fengming Zhu, Fangzhen Lin:
Single-Agent Planning in a Multi-Agent System: A Unified Framework for Type-Based Planners. 2382-2391 - Jiafan Zhuang, Gaofei Han, Zihao Xia, Che Lin, Boxi Wang, Dongliang Wang, Wenji Li, Zhifeng Hao, Ruichu Cai, Zhun Fan:
Robust Policy Learning for Multi-UAV Collision Avoidance with Causal Feature Selection. 2392-2401
Extended Abstracts
- Mohammad Essa Alsomali, Leandro Soriano Marcolino, Barry Porter, Roberto Rodrigues Filho:
Decision-Making in Evolving Environments: A Bayesian Multi-Agent Bandit Framework. 2402-2404 - Ziyan An, Xia Wang, Hendrik Baier, Zirong Chen, Abhishek Dubey, Taylor T. Johnson, Jonathan Sprinkle, Ayan Mukhopadhyay, Meiyi Ma:
Combining LLMs with a Logic-Based Framework to Explain MCTS. 2405-2407 - Vincenzo Auletta, Francesco Carbone, Diodato Ferraioli, Cosimo Vinci:
Adaptive Multi-Round Influence Maximization with Limited Information. 2408-2410 - Ayhan Alp Aydeniz, Enrico Marchesini, Robert Tyler Loftin, Christopher Amato, Kagan Tumer:
Safe Entropic Agents under Team Constraints. 2411-2413 - Haris Aziz, Hau Chan, Xingchen Sha, Toby Walsh, Lirong Xia:
Group Fairness in Multi-period Mobile Facility Location Problems. 2414-2416 - Haris Aziz, Xin Huang, Kei Kimura, Indrajit Saha, Zhaohong Sun, Mashbat Suzuki, Makoto Yokoo:
Weighted Envy-free Allocation with Subsidy. 2417-2419 - Haris Aziz, Grzegorz Lisowski, Mashbat Suzuki, Jeremy Vollen:
Neighborhood Stability in Assignments on Graphs. 2420-2422 - Negar Babashah, Hasti Karimi, Masoud Seddighin, Golnoosh Shahkarami:
On the Distortion of Multi-Winner Elections on the Line Metric. 2423-2425 - Matteo Baldoni, Cristina Baroglio, Stéphane Galland, Roberto Micalizio, Fatma Outay, Stefano Tedeschi:
Interaction Protocols in an Imperative Agent-Oriented Programming Language: the case of BSPL and SARL. 2426-2427 - Marcello Bavaro, Francesco Amigoni:
Multi-Agent Pickup and Delivery with Batteries. 2428-2430 - Curtis Bechtel, Shaddin Dughmi:
Efficient Multi-Agent Delegated Search. 2431-2433 - Tyler J. Becker, Zachary Sunberg:
Bridging the Gap between Partially Observable Stochastic Games and Sparse POMDP Methods. 2434-2436 - Raphaël Berthon, Joost-Pieter Katoen, Munyque Mittelmann, Aniello Murano:
Robust Strategies for Stochastic Multi-Agent Systems. 2437-2439 - Raven Beutner, Bernd Finkbeiner:
Multiplayer Games With Incomplete Information for Hyperproperty Verification. 2440-2442 - Viviane Bonadia dos Santos, Leliane Nunes de Barros, Maria Viviane de Menezes, Silvio do Lago Pereira:
Planning for Temporally Extended Goals based on alpha-CTL. 2443-2445 - Andreas Brännström, Chiaki Sakama, Juan Carlos Nieves:
Formal Verification of Manipulation Dialogues. 2446-2448 - Bugra Çaskurlu, Ali Eser:
(Submodular) Hedonic Games with Common Ranking Property. 2449-2451 - Davide Catta, Angelo Ferrando, Vadim Malvone:
Agreement Games in Multi-Agent Systems. 2452-2454 - Fuguang Chen, Alan Tsang:
The Costly Bargain: Economic Impacts of Price-Seeking Behavior in Aging Populations. 2455-2456 - Haosheng Chen, Yun Hua, Junjie Sheng, Wenhao Li, Bo Jin, Xiangfeng Wang:
Dynamic Conservative Degree Allocation for Offline Multi-Agent Reinforcement Learning. 2457-2459 - Zhixun Chen, Zijing Shi, Yaodong Yang, Meng Fang, Yali Du:
Hierarchical Multi-Agent Framework for Dynamic Macroeconomic Modelling Using Large Language Models. 2460-2462 - Zhuojun Chen, Tacitus Hui, Xinghua Zhu, Dongzhe Su:
Traffic Anomaly Detection through Generative Modeling of Multi-Agent Interactions in Traffic Flow. 2463-2465 - Yukun Cheng, Xiaotie Deng, Baqiao Quan:
Optimal Mechanism Design for Crowdfunding of Public Goods. 2466-2468 - Farès Chouaki, Aurélie Beynier, Nicolas Maudet, Paolo Viappiani:
Fairness in Cooperative Multi-agent Multi-objective Reinforcement Learning using the Expected Scalarized Return. 2469-2471 - Justin Clarke, Przemyslaw Grabowicz, David D. Jensen:
Open-World Classification with Bayesian Gaussian Mixture Models. 2472-2474 - Saar Cohen, Noa Agmon:
Egalitarianism in Online Coalition Formation. 2475-2477 - Ofer Dagan, Tyler J. Becker, Zachary N. Sunberg:
Resolving Multiple-Dynamic Model Uncertainty in Hypothesis-Driven Belief-MDPs. 2478-2480 - Jemma Daniel, Ruan John de Kock, Louay Ben Nessir, Sasha Abramowitz, Omayma Mahjoub, Wiem Khlifi, Juan Claude Formanek, Arnu Pretorius:
Multi-Agent Reinforcement Learning with Selective State-Space Models. 2481-2483 - Koustav De, Palash Dey, Swagato Sanyal:
Voter Participation Control in Online Polls. 2484-2486 - Giuseppe De Giacomo, Yves Lespérance, Gianmarco Parretti, Fabio Patrizi, Renzo Schram:
Managing an Agent's Changing Intentions Using ltlf Synthesis. 2487-2489 - Ruan de Kock, Arnu Pretorius, Jonathan P. Shock:
Is an Exponentially Growing Action Space Really that Bad? Validating a Core Assumption for using Multi-Agent RL. 2490-2492 - Oliver Deane, Oliver Ray:
Symplex: Learning Social Norm Hierarchies by Combining Autonomous Exploration and Expert Imitation. 2493-2495 - Sydney Dolan, Siddharth Nayak, Jasmine Jerry Aloor, Hamsa Balakrishnan:
Asynchronous Cooperative Multi-Agent Reinforcement Learning with Limited Communication. 2496-2498 - Martin Durand, Laurin Erlacher, Johanne Müller Vistisen, Sofia Simola:
Parameterized Complexity of Hedonic Games with Enemy-Oriented Preferences. 2499-2501 - Victoria M. Edwards, M. Ani Hsieh:
Distributed Adaptive Macroscopic Ensemble Task Allocation of Heterogeneous Robot Teams in Dynamic Environments. 2502-2503 - Noga Klein Elmalem, Rica Gonen, Erel Segal-Halevi:
Weighted Envy Freeness With Bounded Subsidies. 2504-2506 - Zeki Doruk Erden, Boi Faltings:
Agential AI for Integrated Continual Learning, Deliberative Behavior, and Comprehensible Models. 2507-2509 - Benjamin Patrick Evans, Sihan Zeng, Sumitra Ganesh, Leo Ardon:
ADAGE: A Generic Two-layer Framework for Adaptive Agent based Modelling. 2510-2513 - Piotr Faliszewski, Lukasz Janeczko, Dusan Knop, Jan Pokorný, Simon Schierreich, Mateusz Sluszniak, Krzysztof Sornat:
Participatory Budgeting Project Strength via Candidate Control. 2514-2516 - Marie Farrell, Angelo Ferrando, Mengwei Xu:
Quantitative Operational Monitoring for BDI Agents. 2517-2519 - Lang Feng, Jiahao Lin, Dong Xing, Li Zhang, De Ma, Gang Pan:
Bidirectional Distillation: A Mixed-Play Framework for Multi-Agent Generalizable Behaviors. 2520-2522 - Grant C. Forbes, Jianxun Wang, Leonardo Villalobos-Arias, Arnav Jhala, David I. Roberts:
Action-Dependent Optimality-Preserving Reward Shaping. 2523-2525 - Kevin Fu, Pierce Howell, Shalin Jain, Harish Ravichandar:
Learning Flexible Heterogeneous Coordination With Capability-Aware Shared Hypernetworks. 2526-2528 - Yuma Fujimoto, Kaito Ariu, Kenshi Abe:
Nash Equilibrium and Learning Dynamics in Three-Player Matching m-Action Games. 2529-2531 - Briti Gangopadhyay, Zhao Wang, Alberto Silvio Chiappa, Shingo Takamatsu:
Adaptive Budget Optimization for Multichannel Advertising Using Combinatorial Bandits. 2532-2534 - Jugal Garg, Thorben Tröbst, Vijay V. Vazirani:
Matching Markets with Chores. 2535-2537 - Madelyn Gatchel, Michael P. Wellman:
Learning Bayesian Game Families, with Application to Mechanism Design. 2538-2540 - Andrea Gatti, Viviana Mascardi, Angelo Ferrando:
ChatBDI: Think BDI, Talk LLM. 2541-2543 - Laurent Gourvès, Michael Lampis, Nikolaos Melissinos, Aris Pagourtzis:
Satisfactory Budget Division. 2544-2546 - Laurent Gourvès, Stefano Moretti, Satya Tamby:
Social Ranking for Feature Selection. 2547-2549 - Bernhard Hilpert, Muhan Hou, Kim Baraka, Joost Broekens:
Can you see how I learn? Human Observers' Inferences about Reinforcement Learning Agents' Learning Processes. 2550-2552 - Niklas Höpner, David Kuric, Herke van Hoof:
Making Universal Policies Universal. 2553-2555 - Shengchao Hu, Wanru Zhao, Weixiong Lin, Li Shen, Ya Zhang, Dacheng Tao:
Prompt Tuning with Diffusion for Few-Shot Pre-trained Policy Generalization. 2556-2558 - Vibulan J, Swapnil Dhamal, Shweta Jain, Ojassvi Kumar, Aman Kumar, Harpreet Singh:
Fair Assignment on Multi-Stage Graphs. 2559-2561 - Nusrath Jahan, Johnathan Mell:
Decoding Negotiation Dynamics: The Impact of Opponent Identity and Privacy on Strategy, Deception, and Emotional Transparency in Human-Agent Interaction. 2562-2564 - Ali Jalal-Kamali, Nikolos Gurney, David V. Pynadath:
Predicting Team Performance from Communications in Simulated Search-and-Rescue. 2565-2567 - Wenzheng Jiang, Ji Wang, Xiongtao Zhang, Weidong Bao, Cheston Tan, Flint Xiaofeng Fan:
FedHPD: Heterogeneous Federated Reinforcement Learning via Policy Distillation. 2568-2570 - Anson Kahng, Joseph Saber:
When to Stop Getting Tested: The Theory of Diagnostic Tests. 2571-2573 - Siva Kailas, Shalin Jain, Harish Ravichandar:
Evaluating and Improving Graph-based Explanation Methods for Multi-Agent Coordination. 2574-2576 - Yasushi Kawase, Bodhayan Roy, Mohammad Azharuddin Sanpui:
Resource Allocation under the Latin Square Constraint. 2577-2578 - Bing-Zhi Ke, Kuang-Da Wang, Wen-Chih Peng:
RallyDiffuser: A Representation-Guided Diffusion Model Framework for Strategic Planning in Badminton. 2579-2581 - Behrad Koohy, Sebastian Stein, Enrico H. Gerding:
Adaptive Microtolling in Competitive Online Congestion Games via Multiagent Reinforcement Learning. 2582-2584 - Krzysztof Kowalczyk, Dominik Baumann, Cristian R. Rojas, Pawel Wachel:
Compensating Latent Nonlinear Dynamics for Practical Consensus Control. 2585-2587 - Mel Krusniak, Hang Xu, Parker Palermo, Forrest Laine:
Online Competitive Information Gathering for Partially Observable Trajectory Games. 2588-2590 - Ashwin Kumar, William Yeoh:
DECAF: Learning to be Fair in Multi-agent Resource Allocation. 2591-2593 - Dattatray Vishnu Kute, Zihao Xu, Yuekang Li, Fethi Rabhi:
Truman: A Large Language Model-based Multi-agent Simulator for Synthetic Money Laundering Data Generation. 2594-2596 - Dmytro Kuzmenko, Nadiya Shvai:
Knowledge Transfer in Model-Based Reinforcement Learning Agents for Efficient Multi-Task Learning. 2597-2599 - Frédérique Lalieu, Tomasz Zurek, Tom M. van Engers:
Model of the Influence of External Signals on the Trust of the Agent in Multi Agent System. 2600-2602 - Hodaya Lampert, Reshef Meir, Kinneret Teodorescu:
To Stand on the Shoulders of Giants: Should We Protect Initial Discoveries in Multi-Agent Exploration? 2603-2605 - Wei-Chen Lee, Alessandro Abate, Michael J. Wooldridge:
Equilibrium Selection via Communication Partition. 2606-2608 - Salomé Lepers, Vincent Thomas, Olivier Buffet:
Observer-Aware Probabilistic Planning under Partial Observability. 2609-2611 - Haorui Li, Jiaqi Liang, Linjing Li, Daniel Zeng:
Offline Meta Reinforcement Learning with Weighted Policy Constraints and Proximal Context Collection. 2612-2614 - Minming Li, Cheng Peng, Ying Wang, Houyu Zhou:
Group-fair Facility Location Games with Externalities. 2615-2617 - Yan Li, Meng Liu, Zhongchen Shi, Yanqing Hou, Liang Xie, Hongbo Chen, Erwei Yin:
Lite-DIO Is Actually What You Need for Efficient Inertial Localization. 2618-2620 - Yaqiao Li, Lata Narayanan, Jaroslav Opatrny, Yi Tian Xu:
Diversity-seeking Swap Games in Networks. 2621-2623 - Zhilin Li, Xiaomei Tao:
Fusing Physical and Cognitive Stimuli: An Eye Movement Emotion Recognition Framework Based on Hierarchical Attention Mechanism. 2624-2626 - Daxin Liu, Vaishak Belle:
What Is a Counterfactual Cause in Action Theories? 2627-2629 - Lunjun Liu, Weilai Jiang, Yaonan Wang:
Tacit Learning with Adaptive Information Selection for Cooperative Multi-Agent Reinforcement Learning. 2630-2632 - Xinjie Liu, Jingqi Li, Filippos Fotiadis, Mustafa O. Karabag, Jesse Milzman, David Fridovich-Keil, Ufuk Topcu:
Policies with Sparse Inter-Agent Dependencies in Dynamic Games: A Dynamic Programming Approach. 2633-2635 - Xu Liu, Tong Yu, Shuai Li:
Adaptive Offline Data Replay in Offline-to-Online Reinforcement Learning. 2636-2638 - Yingzhuo Liu, Shuodi Liu, Hongsong Tang, Yubing Ma, Zikang Li, Junge Zhang, Liuyu Xiang, Zhaofeng He:
RainbowArena: A Multi-Agent Toolkit for Reinforcement Learning and Large Language Models in Competitive Tabletop Games. 2639-2641 - Zeyuan Liu, Kai Yang, Jiafei Lyu, Xiu Li:
CDSA: Conservative Denoising Score-based Algorithm for Offline Reinforcement Learning. 2642-2644 - Panagiotis Lymperopoulos, Vasanth Sarathy:
Tools in the Loop: Quantifying Uncertainty of LLM Question Answering Systems That Use Tools. 2645-2647 - Guoqing Ma, Yuhan Zhang, Yuming Dai, Guangfu Hao, Yang Chen, Shan Yu:
Mitigating Non-Stationarity in Deep Reinforcement Learning with Clustering Orthogonal Weight Modification. 2648-2650 - Mateus G. Machado, Hansenclever F. Bassani:
DyLam: A Dynamic Reward Weighting Framework for Reinforcement Learning Algorithms. 2651-2653 - Yihuan Mao, Yipeng Kang, Peilun Li, Ning Zhang, Wei Xu, Chongjie Zhang:
IBGP: Imperfect Byzantine Generals Problem for Zero-Shot Robustness in Communicative Multi-Agent Systems. 2654-2656 - Carlos Martin, Tuomas Sandholm:
AlphaZeroES: Direct Score Maximization Outperforms Planning Loss Minimization. 2657-2659 - Ryota Maruo, Koh Takeuchi, Hisashi Kashima:
Learning Fair and Preferable Allocations through Neural Network. 2660-2662 - Peta Masters, Daniel Gallagher, Luc Moreau, Mor Vered:
Rethinking Explainable AI: Explanations can be Deceiving. 2663-2665 - Enrique Mateos-Melero, Javier Moralejo-Piñas, Ángela Durán-Pinto, Francisco Martinez-Gil, María Soriano, Fernando Fernández:
Where is the Nearest EV Charging Station? Evolutionary Optimization of the Gas/charging Stations Topology. 2666-2668 - Alexander McCaffrey, Eduardo Alonso, Esther Mondragón:
Predictive Improvement through Latent Space Optimisation. 2669-2671 - Mateus Begnini Melchiades, Gabriel de Oliveira Ramos, Bruno C. da Silva:
Dynamic Option Creation in Option-Critic Reinforcement Learning. 2672-2674 - David Milec, Vojtech Kovarík, Viliam Lisý:
Adapting Beyond the Depth Limit: Counter Strategies in Large Imperfect Information Games. 2675-2677 - Hamza Mohammed, Hang Yin, Sai Chand Boyapati:
Context Adaptive Memory-Efficient LLM Inference for Edge Multi-Agent Systems. 2678-2680 - Jahir Sadik Monon, Deeparghya Dutta Barua, Md. Mosaddek Khan:
Learning Heterogeneous Agent Collaboration in Decentralized Multi-Agent Systems via Intrinsic Motivation. 2681-2683 - Henrik Müller, Daniel Kudenko:
Improving the Effectiveness of Potential-based Reward Shaping in Reinforcement Learning. 2684-2686 - Calarina Muslimani, Bram Grooten, Deepak Ranganatha Sastry Mamillapalli, Mykola Pechenizkiy, Decebal Constantin Mocanu, Matthew E. Taylor:
Boosting Robustness in Preference-Based Reinforcement Learning with Dynamic Sparsity. 2687-2689 - Nasik Muhammad Nafi, William H. Hsu:
A Minimalist Approach to Augmentation-based Self-supervised Representation Learning for On-policy Reinforcement Learning. 2690-2692 - Dung Nguyen, Hung Le, Kien Do, Sunil Gupta, Svetha Venkatesh, Truyen Tran:
Navigating Social Dilemmas with LLM-based Agents via Consideration of Future Consequences. 2693-2695 - Guanyu Nie, Vaneet Aggarwal, Christopher John Quinn:
Stochastic k-Submodular Bandits with Full Bandit Feedback. 2696-2698 - Taylor Olson, Roberto Salas-Damian, Kenneth D. Forbus:
Reasoning and Planning with Dynamic Social Norms. 2699-2701 - Zuzanna Osika, Roxana Radulescu, Jazmin Zatarain Salazar, Frans A. Oliehoek, Pradeep K. Murukannaiah:
Multi-Objective Reinforcement Learning for Water Management. 2702-2704 - Bizhao Pang, Xinting Hu, Mingcheng Zhang, Sameer Alam, Guglielmo Lulli:
Decentralized Deep Reinforcement Learning for Cooperative Multi-Agent Flight Trajectory Planning in Adverse Weather. 2705-2707 - Charly Pecqueux-Guézénec, Stéphane Doncieux, Nicolas Perrin-Gilbert:
Learning to Explore when Mistakes are Not Allowed. 2708-2710 - Arseniy Pertzovsky, Roni Stern, Ariel Felner, Roie Zivan:
Enhancing Lifelong Multi-Agent Path-finding by Using Artificial Potential Fields. 2711-2713 - Luis Pimentel, Sean Ye, James Ellis Grant Pagan, Matthew C. Gombolay:
Diverse Heterogeneous Graph Conditioned Diffusion for Multi-Agent Teaming. 2714-2716 - Gokul Puthumanaillam, Paulo Padrao, Jose Fuentes, Leonardo Bobadilla, Melkior Ornik:
Enhancing Robot Navigation Policies with Task-Specific Uncertainty Management. 2717-2719 - Pranav Rajbhandari, Prithviraj Dasgupta, Donald A. Sofge:
Transformer Guided Coevolution: Improved Team Formation in Multiagent Adversarial Games. 2720-2722 - Rasheed, Parth Nimish Desai, Yash Chaurasia, Sujit Gujar:
Shapley Value-based Approach for Distributing Revenue of Matchmaking of Private Transactions in Blockchains. 2723-2725 - Sebastian Rodriguez, John Thangarajah, Michael Winikoff:
Requirements-based Explainability for Multi Agent Systems. 2726-2728 - Manel Rodriguez-Soto, Nardine Osman, Carles Sierra, Rocio Cintas Garcia, Cristina Farriols Danes, Montserrat Garcia Retortillo, Silvia Minguez Maso, Jordi Martinez Roldan:
Towards Automating the Design of Value-Aligned Clinical Protocols. 2729-2731 - Ryosuke Sato:
Liquid Welfare and Revenue Monotonicity in Adaptive Clinching Auctions. 2732-2734 - Alkmini Sgouritsa, Minas Marios Sotiriou:
On the Existence of EFX Allocations in Multigraphs. 2735-2737 - Liang Shan, Zhengyang Liu, Haoqiang Huang, Zihe Wang:
Environmental Policies within Cournot Oligopoly. 2738-2740 - Junjie Sheng, Wenhao Li, Bo Jin, Hongyuan Zha, Jun Wang, Xiangfeng Wang:
Negotiated Reasoning: On Provably Addressing Relative Over-Generalization. 2741-2743 - Umer Siddique, Peilang Li, Yongcan Cao:
Towards Fair and Efficient Policy Learning in Cooperative Multi-Agent Reinforcement Learning. 2744-2746 - Aditya Vikram Singh, Ethan Rathbun, Emma Graham, Lisa Oakley, Simona Boboila, Peter Chin, Alina Oprea:
Hierarchical Multi-agent Reinforcement Learning for Cyber Network Defense. 2747-2749 - Agamdeep Singh, Jaskirat Singh, P. B. Sujit:
PANDA: Priority-Based Collision Avoidance Framework for Heterogeneous UAVs Navigating in Dense Airspace. 2750-2752 - Ziyang Song, Ziyu Chen, Jinhui Huang, Cheng Zhang, Jingyuan He:
Modeling the Collaborative Edge Data Caching Problem via a Dynamic DCOP. 2753-2755 - Jared Soundy, Mohammad T. Irfan, Hau Chan:
Pure Nash Equilibrium and Strong Nash Equilibrium Computation in Additive Aggregate Games. 2756-2758 - Lennard Sund, Janik Muires, Ramin Ahadi, Konstantina Valogianni, Wolfgang Ketter:
Coordinating Competing Electric Vehicle Fleets: An Agent-Based Charging Capacity Market. 2759-2761 - Abhishek Kumar, Andra Siva Sai Teja, Ganesh Ghalme, Sujit Gujar, Y. Narahari:
Regret Guarantees for a UCB-based Algorithm for Volatile Combinatorial Bandits. 2762-2764 - Lewis Thelen, Vikram Ravindra:
Practical Comparisons of Reservoir Topology Performance and Input Distribution in Digital Reservoir Computers. 2765-2767 - Kyle Tilbury, David Radke:
Dynamic Reward Sharing to Enhance Learning in the Context of Multiagent Teams. 2768-2770 - Aron Vallinder, Edward Hughes:
Cultural Evolution of Cooperation among LLM Agents. 2771-2773 - Guilherme S. Varela, Alberto Sardinha, Francisco S. Melo:
Distributed Value Decomposition Networks with Networked Agents. 2774-2776 - Alicia Vidler, Toby Walsh:
Shifting Power: Leveraging LLMs to Simulate Human Aversion in ABMs of Bilateral Financial Exchanges, A bond market study. 2777-2779 - Mattia Jacopo Villani, Emanuele Natale, Frederik Mallmann-Trenn:
Trading-off Accuracy and Communication Cost in Federated Learning. 2780-2782 - Chulabhaya Wijesundara, Andrea Baisero, Gregory D. Castañón, Alan Carlin, Robert Platt, Christopher Amato:
Leveraging Fully-Observable Solutions for Improved Partially-Observable Offline Reinforcement Learning. 2783-2785 - Richard Willis, Yali Du, Joel Z. Leibo:
Will Systems of LLM Agents Lead to Cooperation: An Investigation into a Social Dilemma. 2786-2788 - Jessica Woodgate, Nirav Ajmeri:
Combining Normative Ethics Principles to Learn Prosocial Behaviour. 2789-2791 - Mingkang Wu, Yongcan Cao:
On-Policy Reinforcement Learning From Failure via Sparse Reward Densification. 2792-2794 - Wannian Xia, Meng Fang, Zihao Guo, Yali Du, Bo Xu:
Integrating Large Language Models with Reinforcement Learning for Generalization in Strategic Card Games. 2795-2797 - Yi Xie, Ziqing Zhou, Chun Ouyang, Siao Liu, Linqiang Hu, Zhongxue Gan:
Heuristics-Assisted Experience Replay Strategy for Cooperative Multi-Agent Reinforcement Learning. 2798-2800 - Tianpei Yang, Srijita Das, Christabel Wayllace, Matthew E. Taylor:
Empowering Generalization for Deep Reinforcement Learning via Symbolic Planning. 2801-2803 - Yi Yang, Tom Holvoet:
Efficient Model Checking with Semantically-Equivalent Models for vGOAL. 2804-2806 - Yue Yang, Bernd Meyer, Frits de Nijs:
Using Assistance Rewards Without Introducing Bias: Overcoming Sparse Rewards in Multi-Agent Reinforcement Learning. 2807-2809 - Zhaohui Yang, Xiaoxuan Wang, Linjing Li:
CPE: A New Paradigm for Policy Extraction in Offline Reinforcement Learning. 2810-2812 - Shiqing Yao, Jiajun Chai, Haixin Yu, Yongzhe Chang, Yuanheng Zhu, Xueqian Wang:
Learning Pre-Trained Tacit Behavior for Efficient Multi-Agent Adversarial Coordination. 2813-2815 - Zixin Ye, Tansu Alpcan, Christopher Leckie:
Local Anomaly Detection with Partial Observation in Multi-agent Systems as a Data Matching Game. 2816-2818 - Haotong Zhang, Wanyuan Wang:
Fast Adaption by Policy Deviation Integral Meta-reinforcement Learning with Applications to High-speed Trains Operation. 2819-2821 - Hengrui Zhang, Youfang Lin, Shuo Shen, Hanfeng Lin, Peng Cheng, Sheng Han, Kai Lv:
Enhancing Offline Safe Reinforcement Learning with Trajectory-Constrained Diffusion Planning. 2822-2824 - Hexiao Zhang, Yanni Tang, Jiamou Liu, Wu Chen:
SFedRec: A Federated Learning Framework for Dynamic Session-based Recommendation. 2825-2828 - Tuo Zhang, Leonardo Stella, Julian Barreiro-Gomez:
Experience-replay Innovative Dynamics. 2829-2831 - Yinuo Zhao, Kun Wu, Tianjiao Yi, Zhiyuan Xu, Zhengping Che, Chi Harold Liu, Jian Tang:
Efficient Training of Generalizable Visuomotor Policies via Control-Aware Augmentation. 2832-2834 - Luis Zhinin-Vera, José J. González-García, Víctor López-Jaquero, Elena Navarro, Pascual González:
Multi-Agent Systems for Bullying Intervention. 2835-2837 - Yihe Zhou, Shunyu Liu, Yunpeng Qing, Tongya Zheng, Kaixuan Chen, Jie Song, Mingli Song:
CADP: Towards Better Centralized Learning for Decentralized Execution in MARL. 2838-2840 - Changxi Zhu, Mehdi Dastani, Shihan Wang:
Reducing Variance Caused by Communication in Decentralized Multi-agent Deep Reinforcement Learning. 2841-2843
Blue Sky Ideas
- Saptarashmi Bandyopadhyay, John (Jack) Cole, Tom Goldstein, David Jacobs:
Multimodal Agentic Model Predictive Control. 2844-2848 - Jeremy Bellay, J. Timothy Balint, Stephen A. Boxwell, Jeffrey Geppert:
Safe Systems with Unsafe Agents: Challenges and Opportunities. 2849-2853 - Virginia Dignum, Loizos Michael, Juan Carlos Nieves, Marija Slavkovik, Julliett Suarez, Andreas Theodorou:
Contesting Black-Box AI Decisions. 2854-2858 - Hermine J. Grosinger:
The Next Level of Long-Term Agent Autonomy - Proactively Acquiring Knowledge and Abilities. 2859-2864 - Reshef Meir:
Tyranny of the Minority in Social Choice: a Call to Arms. 2865-2869 - Yasser Mohammad:
Tackling the Protocol Problem in Automated Negotiation. 2870-2874 - François Olivier, Zied Bouraoui:
Grounding Agent Reasoning in Image Schemas: A Neurosymbolic Approach to Embodied Cognition. 2875-2879 - Abhimanyu Pallavi Sudhir, Long Tran-Thanh:
Market-based Architectures in RL and Beyond. 2880-2884 - Jingyao Ren, Eric Ewing, T. K. Satish Kumar, Sven Koenig, Nora Ayanian:
Empirical Hardness in Multi-Agent Pathfinding: Research Challenges and Opportunities. 2885-2889 - James Rudd-Jones, Mirco Musolesi, María Pérez-Ortiz:
Multi-Agent Reinforcement Learning Simulation for Environmental Policy Synthesis. 2890-2895 - Yingxuan Yang, Qiuying Peng, Jun Wang, Ying Wen, Weinan Zhang:
Unlocking the Potential of Decentralized LLM-based MAS: Privacy Preservation and Monetization in Collective Intelligence. 2896-2900 - Yunfan Zhao, Niclas Boehmer, Aparna Taneja, Milind Tambe:
Towards Foundation-model-based Multiagent System to Accelerate AI for Social Impact. 2901-2907
Doctoral Consortium
- Anastasia Sophia Apeiron:
Responsible Autonomy for Hybrid Intelligence. 2911-2913 - Ayhan Alp Aydeniz:
Learning Diverse Multiagent Behaviors. 2914-2916 - Andrea Baisero:
Role of State in Partially Observable Reinforcement Learning. 2917-2919 - Karen Frilya Celine:
Balancing Fairness and Efficiency in the Allocation of Indivisible Goods. 2920-2922 - Daniel E. Collins:
Human Influences on Decision Making in Multi-Agent Systems. 2923-2925 - Ari Conati:
Collective Decision Making via Automated Reasoning. 2926-2928 - Madelyn Gatchel:
Game-Family Learning for Simulation-Based Games. 2929-2931 - Minghong Geng:
Hierarchical Frameworks for Scaling-up Multi-agent Coordination. 2932-2934 - Everardo Gonzalez:
Influence Based Reward Shaping in Multiagent Systems. 2935-2937 - Victor Guillet:
Extending Consensus-based Task Allocation Algorithms with Bid Intercession to Foster Mixed-Initiative. 2938-2940 - Qishen Han:
Informed Decision-Making via Voting. 2941-2943 - Sylvia S. Kerkhove:
Causality in Multi-Agent Systems. 2944-2946 - Enrique Mateos-Melero:
Efficient Offline Reinforcement Learning Through Dataset Characterization and Reduction. 2950-2952 - Arnau Mayoral-Macau:
Environment-Centered Design of Ethical Environments. 2953-2955 - Zakaria Mehrab:
Modeling and Optimizing Agent-Based Model of Conflict-Induced Forced Migration. 2956-2958 - Daniel Melcer:
Safe Multi-Agent Learning via Shielding in Decentralized Environments. 2959-2961 - Janik Muires:
Agent-Based Modeling of Smart Sustainable Mobility Services, Markets, and Policy. 2965-2967 - Jannik Peters:
Humanlike Emergent Language in Multi-Agent Systems. 2971-2973 - Alexandre S. Pires:
The Impact of Artificial Agents in Human Cooperation Through Indirect Reciprocity. 2974-2976 - Arjun Prakash:
Bi-Level Reinforcement Learning for Multi-Robot Systems. 2977-2978 - Pulkit Rustagi:
Multi-Agent Multi-Objective Planning with Contextual Lexicographic Reward Preferences. 2982-2984 - Lorenzo Serina:
Deep Learning approaches to Goal Recognition. 2985-2987 - Bin Sun:
Different Models for Fair and Efficient Resource Allocation. 2988-2990 - Jessica Woodgate:
Ethical Decision-Making in Multi-Agent Systems. 2991-2993 - Peihong Yu:
Learning with Less Effort: Efficient Training and Generalization in (Multi-)Robot Systems. 2994-2996
Demo Track
- Marco Aruta, Vadim Malvone, Aniello Murano, Vincenzo Pio Palma, Salvatore Romano:
FindMe: A Prototype Videogame AI based on CTL with an Optimized Synthesis Algorithm. 2997-2999 - Reyhan Aydogan, Tim Baarslag, Tamara C. P. Florijn, Katsuhide Fujita, Catholijn M. Jonker, Yasser Mohammad:
[COMP24] The Automated Negotiating Agents Competition (ANAC) 2024 Challenges and Results. 3000-3002 - Hicham Azmani, Andries Rosseau, Marjon Blondeel, Ann Nowé:
A JAX-Accelerated Simulation Framework for Multi-Agent Energy Management in Energy Communities. 3003-3005 - Matteo Baldoni, Samuel Christie, Munindar P. Singh, Amit K. Chopra:
Orpheus: Programming Protocol-Based BDI Agents. 3006-3008 - Julius Gasson, Francesco Belardinelli:
LUNAR: A Runtime Verification Tool for Anomaly Detection in Gas Networks. 3009-3011 - Luigi Bellomarini, Marco Favorito, Giuseppe Galano:
BitML2MCMAS: Strategic Reasoning for Bitcoin Smart Contracts. 3012-3014 - Filippo Bistaffa, Sergio Calo Oliveira:
Recommending Green Routes for Pedestrians to Reduce the Exposure to Air Pollutants in Barcelona. 3015-3017 - Jayati Deshmukh, Zijie Liang, Vahid Yazdanpanah, Sebastian Stein, Sarvapali D. Ramchurn:
Serious Games for Ethical Preference Elicitation. 3018-3020 - Angelo Ferrando, Vadim Malvone:
VITAMIN: VerIficaTion of A MultI ageNt system. 3023-3025 - Xinwei Gao, Arambam James Singh, Gangadhar Royyuru, Michael Yuhas, Arvind Easwaran:
CRLLK: Constrained Reinforcement Learning for Lane Keeping in Autonomous Driving. 3026-3028 - Maya Medjad, Hugo Imbert, Bruno Yun, Raphaël Szymocha, Frédéric Armetta:
Leveraging Graph Structures and Large Language Models for End-to-End Synthetic Task-Oriented Dialogues. 3029-3031 - Tebourbi Hedi, Sana Nouzri, Yazan Mualla, Amro Najjar:
Personalized Language Learning: A Multi-Agent System Leveraging LLMs for Teaching Luxembourgish. 3032-3034 - Zhiwei (Tony) Qin, Jianming Zhou:
Eva: An LLM-based Multilingual Voice-agent Network for Restaurant Operations. 3035-3037 - David Radke, Kyle Tilbury:
Simulating Tracking Data to Advance Sports Analytics Research. 3038-3040 - Vítor Crista, Diogo Martinho, Goreti Marreiros:
Chat4Elderly: A Multi-Agent System for Personalized Wellness Using Generative AI and Wearable Technology. 3041-3043 - Simon Rey, Ulle Endriss:
The Game Academy: Learn while playing, and play while learning! 3044-3046 - Kenneth See, Nicholas MacGregor Garcia, Xiaofan Li:
Simulating Blockchain Applications in Large-Value Payment Systems through Agent-Based Modeling. 3047-3049 - Kivanç Serefoglu, Önder Gürcan, Reyhan Aydogan:
UAV Marketplace Simulation Tool for BVLOS Operations. 3050-3052 - Chathurangi Shyalika, Renjith Prasad, Alaa T. Al Ghazo, Darssan Eswaramoorthi, Sara Shree Muthuselvam, Amit P. Sheth:
SmartPilot: Agent-Based CoPilot for Intelligent Manufacturing. 3053-3055 - Markus Utke, Simon Rey, Ulle Endriss:
Pabuviz.org: A Visualisation Platform to Explore Participatory Budgeting Elections. 3056-3058 - Martin Weiss, Nasim Rahaman, Chris Pal:
MapBot: A Multi-Modal Agent for Geospatial Analysis. 3059-3061 - Peijie Zhao, Zunayed Arefin, Felipe Meneguzzi, Ramon Fraga Pereira:
Intention Recognition in Real-Time Interactive Navigation Maps. 3062-3064
JAAMAS Track
- Edmond Awad, Sydney Levine, Andrea Loreggia, Nicholas Mattei, Iyad Rahwan, Francesca Rossi, Kartik Talamadupula, Joshua B. Tenenbaum, Max Kleiman-Weiner:
When Is It Acceptable to Break the Rules? Knowledge Representation of Moral Judgements Based on Empirical Data (Extended Abstract). 3065-3067 - Jake Barrett, Kobi Gal, Loizos Michael, Dan Vilenchik:
Beyond the Echo Chamber: Modelling Open-Mindedness in Citizens' Assemblies. 3068-3070 - Priel Levy, Yonatan Aumann, David Sarne:
Contest Partitioning in Binary Contests: Costly, yet Beneficial. 3071-3073 - Dimitris Michailidis, Mayesha Tasnim, Sennay Ghebreab, Fernando P. Santos:
A summary of: Tackling School Segregation with Transportation Network Interventions - An Agent-Based Modelling Approach. 3074-3076 - Simon Rey, Ulle Endriss:
Epistemic Selection of Costly Alternatives: The Case of Participatory Budgeting (Extended Abstract). 3077-3079 - Mayesha Tasnim, Youri Weesie, Sennay Ghebreab, Max Baak:
Strategic Manipulation of Preferences in the Rank Minimization Mechanism. 3080-3082 - Qinghao Wang, Yaodong Yang:
Carbon Trading Supply Chain Management Based on Constrained Deep Reinforcement Learning. 3083-3086 - Michael P. Wellman, Katherine Mayo:
Navigating in a Space of Game Views (extended abstract). 3087-3088 - Richard Willis, Yali Du, Joel Z. Leibo, Michael Luck:
Resolving Social Dilemmas with Minimal Reward Transfer - Extended Abstract. 3089-3091
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.