default search action
23rd AAMAS 2024: Auckland, New Zealand
- Mehdi Dastani, Jaime Simão Sichman, Natasha Alechina, Virginia Dignum:
Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2024, Auckland, New Zealand, May 6-10, 2024. International Foundation for Autonomous Agents and Multiagent Systems / ACM 2024
Keynote Talks
- Ann Nowé:
Trustworthy Reinforcement Learning: Opportunities and Challenges. 1 - Liz Sonenberg:
Agents and Humans: Trajectories and Perspectives. 2 - Michael Winikoff:
30 Years of Engineering Multi-Agent Systems: What and Why? 3
Full Research Paper
- Sami Abuhaimed, Sandip Sen:
Team Performance and User Satisfaction in Mixed Human-Agent Teams. 4-12 - Abhijin Adiga, Yohai Trabelsi, Tanvir Ferdousi, Madhav V. Marathe, S. S. Ravi, Samarth Swarup, Anil Kumar S. Vullikanti, Mandy L. Wilson, Sarit Kraus, Reetwika Basu, Supriya Savalkar, Matthew Yourek, Michael Brady, Kirti Rajagopalan, Jonathan Yoder:
Value-based Resource Matching with Fairness Criteria: Application to Agricultural Water Trading. 13-21 - Alba Aguilera, Nieves Montes, Georgina Curto, Carles Sierra, Nardine Osman:
Can Poverty Be Reduced by Acting on Discrimination? An Agent-based Model for Policy Making. 22-30 - Pragnya Alatur, Giorgia Ramponi, Niao He, Andreas Krause:
Provably Learning Nash Policies in Constrained Markov Potential Games. 31-39 - Evan Albers, Mohammad T. Irfan, Matthew J. Bosch:
Beliefs, Shocks, and the Emergence of Roles in Asset Markets: An Agent-Based Modeling Approach. 40-48 - Georgios Amanatidis, Aris Filos-Ratsikas, Philip Lazos, Evangelos Markakis, Georgios Papasotiropoulos:
On the Potential and Limitations of Proxy Voting: Delegation with Incomplete Votes. 49-57 - Giorgio Angelotti, Caroline P. C. Chanel, Adam Henrique Moreira Pinto, Christophe Lounis, Corentin Chauffaut, Nicolas Drougard:
Offline Risk-sensitive RL with Partial Observability to Enhance Performance in Human-Robot Teaming. 58-67 - Nemanja Antonic, Raina Zakir, Marco Dorigo, Andreagiovanni Reina:
Collective Robustness of Heterogeneous Decision-Makers Against Stubborn Individuals. 68-77 - Thomas Archbold, Bart de Keijzer, Carmine Ventre:
Willy Wonka Mechanisms. 78-86 - Gennaro Auricchio, Jie Zhang, Mengxiao Zhang:
Extended Ranking Mechanisms for the m-Capacitated Facility Location Problem in Bayesian Mechanism Design. 87-95 - Shaojie Bai, Dongxia Wang, Tim Muller, Peng Cheng, Jiming Chen:
Stability of Weighted Majority Voting under Estimated Weights. 96-104 - James P. Bailey, Craig A. Tovey:
Impact of Tie-Breaking on the Manipulability of Elections. 105-113 - Daniel Bairamian, Philippe Marcotte, Joshua Romoff, Gabriel Robert, Derek Nowrouzezahrai:
Minimax Exploiter: A Data Efficient Approach for Competitive Self-Play. 114-122 - Gabriel Ballot, Vadim Malvone, Jean Leneutre, Youssef Laarouchi:
Strategic Reasoning under Capacity-constrained Agents. 123-131 - Arti Bandhana, Tomás Kroupa, Sebastián García:
Trust in Shapley: A Cooperative Quest for Global Trust in P2P Network. 132-140 - Paul Barde, Jakob Foerster, Derek Nowrouzezahrai, Amy Zhang:
A Model-Based Solution to the Offline Multi-Agent Reinforcement Learning Coordination Problem. 141-150 - Siddharth Barman, Debajyoti Kar, Shraddha Pathak:
Parameterized Guarantees for Almost Envy-Free Allocations. 151-159 - Francesco Belardinelli, Wojtek Jamroga, Munyque Mittelmann, Aniello Murano:
Verification of Stochastic Multi-Agent Systems with Forgetful Strategies. 160-169 - Michael Bernreiter, Jan Maly, Oliviero Nardi, Stefan Woltran:
Combining Voting and Abstract Argumentation to Understand Online Discussions. 170-179 - Raven Beutner, Bernd Finkbeiner, Hadar Frenkel, Niklas Metzger:
Monitoring Second-Order Hyperproperties. 180-188 - Raven Beutner, Bernd Finkbeiner:
Hyper Strategy Logic. 189-197 - Rangeet Bhattacharyya, Parvik Dave, Palash Dey, Swaprava Nath:
Optimal Referral Auction Design. 198-206 - Vittorio Bilò, Michele Flammini, Gianpiero Monaco, Luca Moscardelli, Cosimo Vinci:
On Green Sustainability of Resource Selection Games with Equitable Cost-Sharing. 207-215 - Natasa Bolic, Tommaso Cesari, Roberto Colomboni:
An Online Learning Theory of Brokerage. 216-224 - Martin Bullinger, Rohith Reddy Gangam, Parnian Shahkar:
Robust Popular Matchings. 225-233 - Turgay Caglar, Sarath Sreedharan:
HELP! Providing Proactive Support in the Presence of Knowledge Asymmetry. 234-243 - Ioannis Caragiannis, Kristoffer Arnsfelt Hansen, Nidhi Rathi:
On the Complexity of Pareto-Optimal and Envy-Free Lotteries. 244-252 - Alessandro Carminati, Davide Azzalini, Simone Vantini, Francesco Amigoni:
A Distributed Approach for Fault Detection in Swarms of Robots. 253-261 - Matteo Castiglioni, Alberto Latino, Alberto Marchesi, Giulia Romano, Nicola Gatti, Chokha Palayamkottai:
Finding Effective Ad Allocations: How to Exploit User History. 262-270 - Davide Catta, Jean Leneutre, Vadim Malvone, Aniello Murano:
Obstruction Alternating-time Temporal Logic: A Strategic Logic to Reason about Dynamic Models. 271-280 - Jiajun Chai, Yuqian Fu, Dongbin Zhao, Yuanheng Zhu:
Aligning Credit for Multi-Agent Cooperation via Model-based Counterfactual Imagination. 281-289 - Jaël Champagne Gareau, Marc-André Lavoie, Guillaume Gosset, Éric Beaudry:
Cooperative Electric Vehicles Planning. 290-298 - Yashovardhan S. Chati, Ramasubramanian Suriyanarayanan, Arunchandar Vasan:
Think Global, Act Local - Agent-Based Inline Recovery for Airline Operations. 299-307 - Chao Chen, Dawei Wang, Feng Mao, Jiacheng Xu, Zongzhang Zhang, Yang Yu:
Deep Anomaly Detection via Active Anomaly Search. 308-316 - Ruifeng Chen, Xu-Hui Liu, Tian-Shuo Liu, Shengyi Jiang, Feng Xu, Yang Yu:
Foresight Distribution Adjustment for Off-policy Reinforcement Learning. 317-325 - Weiqin Chen, James Onyejizu, Long Vu, Lan Hoang, Dharmashankar Subramanian, Koushik Kar, Sandipan Mishra, Santiago Paternain:
Adaptive Primal-Dual Method for Safe Reinforcement Learning. 326-334 - Yuhui Chen, Haoran Li, Dongbin Zhao:
Boosting Continuous Control with Consistency Policy. 335-344 - Chen Cheng, Jinglai Li:
ODEs Learn to Walk: ODE-Net based Data-Driven Modeling for Crowd Dynamics. 345-353 - Mattia Chiari, Alfonso Emilio Gerevini, Andrea Loreggia, Luca Putelli, Ivan Serina:
Fast and Slow Goal Recognition. 354-362 - Dmitry Chistikov, Luisa Estrada, Mike Paterson, Paolo Turrini:
Learning a Social Network by Influencing Opinions. 363-371 - Sung-Ho Cho, Kei Kimura, Kiki Liu, Kwei-guu Liu, Zhengjie Liu, Zhaohong Sun, Kentaro Yahiro, Makoto Yokoo:
Fairness and Efficiency Trade-off in Two-sided Matching. 372-380 - Ayush Chopra, Arnau Quera-Bofarull, Nurullah Giray Kuru, Michael J. Wooldridge, Ramesh Raskar:
Private Agent-Based Modeling. 381-390 - Ayush Chopra, Jayakumar Subramanian, Balaji Krishnamurthy, Ramesh Raskar:
flame: A Framework for Learning in Agent-based ModEls. 391-399 - Clarissa Costen, Anna Gautier, Nick Hawes, Bruno Lacerda:
Multi-Robot Allocation of Assistance from a Shared Uncertain Operator. 400-408 - Gergely Csáji:
A Simple 1.5-approximation Algorithm for a Wide Range of Maximum Size Stable Matching Problems. 409-415 - Sankarshan Damle, Manisha Padala, Sujit Gujar:
Designing Redistribution Mechanisms for Reducing Transaction Fees in Blockchains. 416-424 - Argyrios Deligkas, Eduard Eiben, Tiger-Lily Goldsmith:
The Parameterized Complexity of Welfare Guarantees in Schelling Segregation. 425-433 - Davide Dell'Anna, Pradeep K. Murukannaiah, Bernd Dudzik, Davide Grossi, Catholijn M. Jonker, Catharine Oertel, Pinar Yolum:
Toward a Quality Model for Hybrid Intelligence Teams. 434-443 - Rati Devidze, Parameswaran Kamalaruban, Adish Singla:
Informativeness of Reward Functions in Reinforcement Learning. 444-452 - Sheelabhadra Dey, James Ault, Guni Sharon:
Continual Optimistic Initialization for Value-Based Reinforcement Learning. 453-462 - Jack Dippel, Max Dupré la Tour, April Niu, Sanjukta Roy, Adrian Vetta:
Gerrymandering Planar Graphs. 463-471 - Matheus Aparecido do Carmo Alves, Amokh Varma, Yehia Elkhatib, Leandro Soriano Marcolino:
It Is Among Us: Identifying Adversaries in Ad-hoc Domains using Q-valued Bayesian Estimations. 472-480 - Vitaliy Dolgorukov, Rustam Galimullin, Maksim Gladyshev:
Dynamic Epistemic Logic of Resource Bounded Information Mining Agents. 481-489 - Joel Dyer, Arnau Quera-Bofarull, Nicholas Bishop, J. Doyne Farmer, Anisoara Calinescu, Michael J. Wooldridge:
Population Synthesis as Scenario Generation for Simulation-based Planning under Uncertainty. 490-498 - Soroush Ebadian, Aris Filos-Ratsikas, Mohamad Latifian, Nisarg Shah:
Computational Aspects of Distortion. 499-507 - Taha Eghtesad, Sirui Li, Yevgeniy Vorobeychik, Aron Laszka:
Multi-Agent Reinforcement Learning for Assessing False-Data Injection Attacks on Transportation Networks. 508-515 - Hannes Eriksson, Tommy Tram, Debabrota Basu, Mina Alibeigi, Christos Dimitrakakis:
Reinforcement Learning in the Wild with Maximum Likelihood-based Model Transfer. 516-524 - Ahmad Esmaeili, Zahra Ghorrati, Eric T. Matson:
Holonic Learning: A Flexible Agent-based Distributed Machine Learning Framework. 525-533 - Benjamin Patrick Evans, Sumitra Ganesh:
Learning and Calibrating Heterogeneous Bounded Rational Market Behaviour with Multi-agent Reinforcement Learning. 534-543 - Amy Fang, Hadas Kress-Gazit:
High-Level, Collaborative Task Planning Grammar and Execution for Heterogeneous Agents. 544-552 - Jiazhu Fang, Wenjing Liu:
Facility Location Games with Fractional Preferences and Limited Resources. 553-561 - Liangda Fang, Meihong Yang, Dingliang Cheng, Yunlai Hao, Quanlong Guan, Liping Xiong:
Generalized Strategy Synthesis of Infinite-state Impartial Combinatorial Games via Exact Binary Classification. 562-570 - Qihui Feng, Gerhard Lakemeyer:
Probabilistic Multi-agent Only-Believing. 571-579 - Benedetta Flammini, Davide Azzalini, Francesco Amigoni:
Preventing Deadlocks for Multi-Agent Pickup and Delivery in Dynamic Environments. 580-588 - Grant C. Forbes, Nitish Gupta, Leonardo Villalobos-Arias, Colin M. Potts, Arnav Jhala, David L. Roberts:
Potential-Based Reward Shaping for Intrinsic Motivation. 589-597 - Elliot Fosong, Arrasy Rahman, Ignacio Carlucho, Stefano V. Albrecht:
Learning Complex Teamwork Tasks using a Given Sub-task Decomposition. 598-606 - Kipp McAdam Freud, Nathan F. Lepora, Matt W. Jones, Cian O'Donnell:
BrainSLAM: SLAM on Neural Population Activity Data. 607-613 - Tobias Friedrich, Andreas Göbel, Nicolas Klodt, Martin S. Krejca, Marcus Pappik:
From Market Saturation to Social Reinforcement: Understanding the Impact of Non-Linearity in Information Diffusion Models. 614-622 - Kitty Fung, Qizhen Zhang, Chris Lu, Jia Wan, Timon Willi, Jakob N. Foerster:
Analysing the Sample Complexity of Opponent Shaping. 623-631 - Jasmina Gajcin, Ivana Dusparic:
RACCER: Towards Reachable and Certain Counterfactual Explanations for Reinforcement Learning. 632-640 - Daniel Garces, Stephanie Gil:
Surge Routing: Event-informed Multiagent Reinforcement Learning for Autonomous Rideshare. 641-650 - Yaoxin Ge, Yao Zhang, Dengji Zhao, Zhihao Gavin Tang, Hu Fu, Pinyan Lu:
Incentives for Early Arrival in Cooperative Games. 651-659 - Zemuy Tesfay Gebrekidan, Sebastian Stein, Timothy J. Norman:
Deep Reinforcement Learning with Coalition Action Selection for Online Combinatorial Resource Allocation with Arbitrary Action Space. 660-668 - Ian Gemp, Marc Lanctot, Luke Marris, Yiran Mao, Edgar A. Duéñez-Guzmán, Sarah Perrin, Andras Gyorgy, Romuald Elie, Georgios Piliouras, Michael Kaisers, Daniel Hennes, Kalesha Bullard, Kate Larson, Yoram Bachrach:
Approximating the Core via Iterative Coalition Sampling. 669-678 - Farnoud Ghasemi, Rafal Kucharski:
Modelling the Rise and Fall of Two-sided Markets. 679-687 - Shivam Goel, Yichen Wei, Panagiotis Lymperopoulos, Klára Churá, Matthias Scheutz, Jivko Sinapov:
NovelGym: A Flexible Ecosystem for Hybrid Planning and Learning Agents Designed for Open Worlds. 688-696 - Salil Gokhale, Samarth Singla, Shivika Narang, Rohit Vaish:
Capacity Modification in the Stable Matching Problem. 697-705 - Laurent Gourvès, Gianpiero Monaco:
Nash Stability in Hedonic Skill Games. 706-714 - Moritz Graf, Thorsten Engesser, Bernhard Nebel:
Symbolic Computation of Sequential Equilibria. 715-723 - Sven Gronauer, Tom Haider, Felippe Schmoeller da Roza, Klaus Diepold:
Reinforcement Learning with Ensemble Model Predictive Safety Certification. 724-732 - Bram Grooten, Tristan Tomilin, Gautham Vasan, Matthew E. Taylor, A. Rupam Mahmood, Meng Fang, Mykola Pechenizkiy, Decebal Constantin Mocanu:
MaDi: Learning to Mask Distractions for Generalization in Visual Deep Reinforcement Learning. 733-742 - Cong Guan, Ruiqi Xue, Ziqian Zhang, Lihe Li, Yi-Chen Li, Lei Yuan, Yang Yu:
Cost-aware Offline Safe Meta Reinforcement Learning with Robust In-Distribution Online Task Adaptation. 743-751 - Hao Guo, Zhen Wang, Junliang Xing, Pin Tao, Yuanchun Shi:
Cooperation and Coordination in Heterogeneous Populations with Interaction Diversity. 752-760 - Gauri Gupta, Ritvik Kapila, Ayush Chopra, Ramesh Raskar:
First 100 days of Pandemic: An Interplay of Pharmaceutical, Behavioral and Digital Interventions - A Study using Agent Based Modeling. 761-770 - Balint Gyevnar, Cheng Wang, Christopher G. Lucas, Shay B. Cohen, Stefano V. Albrecht:
Causal Explanations for Sequential Decision-Making in Multi-Agent Systems. 771-779 - Vishwa Prakash HV, Prajakta Nimbhorkar:
Weighted Proportional Allocations of Indivisible Goods and Chores: Insights via Matchings. 780-788 - Hairi, Zifan Zhang, Jia Liu:
Sample and Communication Efficient Fully Decentralized MARL Policy Evaluation via a New Approach: Local TD Update. 789-797 - Chaeeun Han, Jose Paolo Talusan, Daniel Freudberg, Ayan Mukhopadhyay, Abhishek Dubey, Aron Laszka:
Forecasting and Mitigating Disruptions in Public Bus Transit Services. 798-806 - Yifan He, Abdallah Saffidine, Michael Thielscher:
Solving Two-player Games with QBF Solvers in General Game Playing. 807-815 - Yu He, Alexander Lam, Minming Li:
Facility Location Games with Scaling Effects. 816-824 - Hadi Hosseini, Andrew McGregor, Rik Sengupta, Rohit Vaish, Vignesh Viswanathan:
Tight Approximations for Graphical House Allocation. 825-833 - Tianyi Hu, Zhiqiang Pu, Xiaolin Ai, Tenghai Qiu, Jianqiang Yi:
Measuring Policy Distance for Multi-Agent Reinforcement Learning. 834-842 - Yudong Hu, Congying Han, Tiande Guo, Hao Xiao:
Applying Opponent Modeling for Automatic Bidding in Online Repeated Auctions. 843-851 - Halvard Hummel, Ayumi Igarashi:
Keeping the Harmony Between Neighbors: Local Fairness in Graph Fair Division. 852-860 - Aamal Abbas Hussain, Dan Leonte, Francesco Belardinelli, Georgios Piliouras:
On the Stability of Learning in Network Games with Many Players. 861-870 - David Hyland, Julian Gutierrez, Shankaranarayanan Krishna, Michael J. Wooldridge:
Rational Verification with Quantitative Probabilistic Goals. 871-879 - Alexandre Yukio Ichida, Felipe Meneguzzi, Rafael C. Cardoso:
BDI Agents in Natural Language Environments. 880-888 - Chikadibia Ihejimba, Rym Z. Wenkstern:
A Cloud-Based Microservices Solution for Multi-Agent Traffic Control Systems. 889-897 - Tatsuya Iwase, Aurélie Beynier, Nicolas Bredèche, Nicolas Maudet, Jason R. Marden:
Is Limited Information Enough? An Approximate Multi-agent Coverage Control in Non-Convex Discrete Environments. 898-906 - Saïd Jabbour, Yue Ma, Badran Raddaoui:
Towards a Principle-based Framework for Repair Selection in Inconsistent Knowledge Bases. 907-915 - Nusrath Jahan, Johnathan Mell:
Unraveling the Tapestry of Deception and Personality: A Deep Dive into Multi-Issue Human-Agent Negotiation Dynamics. 916-925 - Wojciech Jamroga, Munyque Mittelmann, Aniello Murano, Giuseppe Perelli:
Playing Quantitative Games Against an Authority: On the Module Checking Problem. 926-934 - Lukasz Janeczko, Jérôme Lang, Grzegorz Lisowski, Stanislaw Szufa:
Discovering Consistent Subelections. 935-943 - Chengxing Jia, Fuxiang Zhang, Yi-Chen Li, Chenxiao Gao, Xu-Hui Liu, Lei Yuan, Zongzhang Zhang, Yang Yu:
Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation. 944-953 - Junqi Jiang, Francesco Leofante, Antonio Rago, Francesca Toni:
Recourse under Model Multiplicity via Argumentative Ensembling. 954-963 - Philip Jordan, Florian Grötschla, Flint Xiaofeng Fan, Roger Wattenhofer:
Decentralized Federated Policy Gradient with Byzantine Fault-Tolerance and Provably Fast Convergence. 964-972 - Matej Jusup, Barna Pásztor, Tadeusz Janik, Kenan Zhang, Francesco Corman, Andreas Krause, Ilija Bogunovic:
Safe Model-Based Multi-Agent Mean-Field Reinforcement Learning. 973-982 - Jonas Karge, Juliette-Michelle Burkhardt, Sebastian Rudolph, Dominik Rusovac:
To Lead or to be Led: A Generalized Condorcet Jury Theorem under Dependence. 983-991 - Stanislaw Kazmierowski, Marcin Dziubinski:
Efficient Method for Finding Optimal Strategies in Chopstick Auctions with Uniform Objects Values. 992-1000 - Akbir Khan, Timon Willi, Newton Kwan, Andrea Tacchetti, Chris Lu, Edward Grefenstette, Tim Rocktäschel, Jakob N. Foerster:
Scaling Opponent Shaping to High Dimensional Games. 1001-1010 - Oz Kilic, Alan Tsang:
Catfished! Impacts of Strategic Misrepresentation in Online Dating. 1011-1019 - Aleksei Y. Kondratev, Egor Ianovski:
Veto Core Consistent Preference Aggregation. 1020-1028 - Matthias Köppe, Martin Koutecký, Krzysztof Sornat, Nimrod Talmon:
Fine-Grained Liquid Democracy for Cumulative Ballots. 1029-1037 - Daniel Koyfman, Shahaf S. Shperberg, Dor Atzmon, Ariel Felner:
Minimizing State Exploration While Searching Graphs with Unknown Obstacles. 1038-1046 - Kalle Kujanpää, Amin Babadi, Yi Zhao, Juho Kannala, Alexander Ilin, Joni Pajarinen:
Continuous Monte Carlo Graph Search. 1047-1056 - Pooja Kulkarni, Rucha Kulkarni, Ruta Mehta:
Approximating APS Under Submodular and XOS Valuations with Binary Marginals. 1057-1065 - Otto Kuusela, Debraj Roy:
Higher Order Reasoning under Intent Uncertainty Reinforces the Hobbesian Trap. 1066-1074 - Alexander Lam, Haris Aziz, Bo Li, Fahimeh Ramezani, Toby Walsh:
Proportional Fairness in Obnoxious Facility Location. 1075-1083 - Hung Le, Kien Do, Dung Nguyen, Svetha Venkatesh:
Beyond Surprise: Improving Exploration Through Surprise Novelty. 1084-1092 - Linh Le Pham Van, Hung The Tran, Sunil Gupta:
Policy Learning for Off-Dynamics RL with Deficient Support. 1093-1100 - Chin-wing Leung, Shuyue Hu, Ho-fung Leung:
The Stochastic Evolutionary Dynamics of Softmax Policy Gradient in Games. 1101-1109 - Chin-wing Leung, Paolo Turrini:
Learning Partner Selection Rules that Sustain Cooperation in Social Dilemmas with the Option of Opting Out. 1110-1118 - Chaya Levinger, Noam Hazon, Sofia Simola, Amos Azaria:
Coalition Formation with Bounded Coalition Size. 1119-1127 - Bo Li, Ankang Sun, Shiji Xing:
Bounding the Incentive Ratio of the Probabilistic Serial Rule. 1128-1136 - Lu Li, Jiafei Lyu, Guozheng Ma, Zilin Wang, Zhenjie Yang, Xiu Li, Zhiheng Li:
Normalization Enhances Generalization in Visual Reinforcement Learning. 1137-1146 - Pengdeng Li, Shuxin Li, Xinrun Wang, Jakub Cerný, Youzhi Zhang, Stephen McAleer, Hau Chan, Bo An:
Grasper: A Generalist Pursuer for Pursuit-Evasion Problems. 1147-1155 - Xinran Li, Jun Zhang:
Context-aware Communication for Multi-agent Reinforcement Learning. 1156-1164 - Yixuan Li, Wanyuan Wang, Weiyi Xu, Yanchen Deng, Weiwei Wu:
Factor Graph Neural Network Meets Max-Sum: A Real-Time Route Planning Algorithm for Massive-Scale Trips. 1165-1173 - Zhenglong Li, Vincent W. L. Tam, Kwan L. Yeung:
Developing a Multi-agent and Self-adaptive Framework with Deep Reinforcement Learning for Dynamic Portfolio Risk Management. 1174-1182 - Zihao Li, Shengxin Liu, Xinhang Lu, Biaoshuai Tao, Yichen Tao:
A Complete Landscape for the Price of Envy-Freeness. 1183-1191 - Dayang Liang, Yaru Zhang, Yunlong Liu:
Episodic Reinforcement Learning with Expanded State-reward Space. 1192-1200 - Qian Lin, Chao Yu, Zongkai Liu, Zifan Wu:
Policy-regularized Offline Multi-objective Reinforcement Learning. 1201-1209 - Daxin Liu, Vaishak Belle:
Progression with Probabilities in the Situation Calculus: Representation and Succinctness. 1210-1218 - Jijia Liu, Chao Yu, Jiaxuan Gao, Yuqing Xie, Qingmin Liao, Yi Wu, Yu Wang:
LLM-Powered Hierarchical Language Agent for Real-time Human-AI Coordination. 1219-1228 - Jinyi Liu, Yi Ma, Jianye Hao, Yujing Hu, Yan Zheng, Tangjie Lv, Changjie Fan:
A Trajectory Perspective on the Role of Data Sampling Techniques in Offline Reinforcement Learning. 1229-1237 - Qidong Liu, Chaoyue Liu, Shaoyao Niu, Cheng Long, Jie Zhang, Mingliang Xu:
2D-Ptr: 2D Array Pointer Network for Solving the Heterogeneous Capacitated Vehicle Routing Problem. 1238-1246 - Siqi Liu, Luke Marris, Marc Lanctot, Georgios Piliouras, Joel Z. Leibo, Nicolas Heess:
Neural Population Learning beyond Symmetric Zero-Sum Games. 1247-1255 - Ziqi Liu, Laurence Liu:
GraphSAID: Graph Sampling via Attention based Integer Programming Method. 1256-1264 - Robert T. Loftin, Mustafa Mert Çelikok, Herke van Hoof, Samuel Kaski, Frans A. Oliehoek:
Uncoupled Learning of Differential Stackelberg Equilibria with Commitments. 1265-1273 - Xingzhou Lou, Junge Zhang, Ziyan Wang, Kaiqi Huang, Yali Du:
Safe Reinforcement Learning with Free-form Natural Language Constraints and Pre-Trained Language Models. 1274-1282 - Jiaming Lu, Jingqing Ruan, Haoyuan Jiang, Ziyue Li, Hangyu Mao, Rui Zhao:
DuaLight: Enhancing Traffic Signal Control by Leveraging Scenario-Specific and Scenario-Shared Knowledge. 1283-1291 - Xinpeng Lu, Heng Song, Huailing Ma, Junwu Zhu:
A Task-Driven Multi-UAV Coalition Formation Mechanism. 1292-1300 - Baiting Luo, Yunuo Zhang, Abhishek Dubey, Ayan Mukhopadhyay:
Act as You Learn: Adaptive Decision-Making in Non-Stationary Markov Decision Processes. 1301-1309 - Panagiotis Lymperopoulos, Matthias Scheutz:
Oh, Now I See What You Want: Learning Agent Models with Internal States from Observations. 1310-1318 - Haoxiang Ma, Chongyang Shi, Shuo Han, Michael R. Dorothy, Jie Fu:
Covert Planning aganist Imperfect Observers. 1319-1327 - Haozhe Ma, Thanh Vinh Vo, Tze-Yun Leong:
Mixed-Initiative Bayesian Sub-Goal Optimization in Hierarchical Reinforcement Learning. 1328-1336 - Shivakumar Mahesh, Anshuka Rangi, Haifeng Xu, Long Tran-Thanh:
Attacking Multi-Player Bandits and How to Robustify Them. 1337-1345 - Saaduddin Mahmud, Marcell Vazquez-Chanlatte, Stefan J. Witwicki, Shlomo Zilberstein:
Explaining the Behavior of POMDP-based Agents Through the Impact of Counterfactual Information. 1346-1354 - Fumiyasu Makinoshima, Tetsuro Takahashi, Yusuke Oishi:
Bayesian Behavioural Model Estimation for Live Crowd Simulation. 1355-1362 - Hangyu Mao, Rui Zhao, Ziyue Li, Zhiwei Xu, Hao Chen, Yiqun Chen, Bin Zhang, Zhen Xiao, Junge Zhang, Jiangjin Yin:
PDiT: Interleaving Perception and Decision-making Transformers for Deep Reinforcement Learning. 1363-1371 - Zakaria Mehrab, Logan Stundal, Samarth Swarup, Srinivasan Venkatramanan, Bryan Lewis, Henning S. Mortveit, Christopher L. Barrett, Abhishek Pandey, Chad R. Wells, Alison P. Galvani, Burton H. Singer, David Leblang, Rita R. Colwell, Madhav V. Marathe:
Network Agency: An Agent-based Model of Forced Migration from Ukraine. 1372-1380 - Michela Meister, Jon M. Kleinberg:
Containing the Spread of a Contagion on a Tree. 1381-1389 - Qirui Mi, Siyu Xia, Yan Song, Haifeng Zhang, Shenghao Zhu, Jun Wang:
TaxAI: A Dynamic Economic Simulator and Benchmark for Multi-agent Reinforcement Learning. 1390-1399 - Adway Mitra, Palash Dey:
Evaluating District-based Election Surveys with Synthetic Dirichlet Likelihood. 1400-1408 - Shuwa Miura, Shlomo Zilberstein:
Observer-Aware Planning with Implicit and Explicit Communication. 1409-1417 - Zhaobin Mo, Yongjie Fu, Xuan Di:
PI-NeuGODE: Physics-Informed Graph Neural Ordinary Differential Equations for Spatiotemporal Trajectory Prediction. 1418-1426 - Nasik Muhammad Nafi, Raja Farrukh Ali, William H. Hsu, Kevin Duong, Mason Vick:
Policy Optimization using Horizon Regularized Advantage to Improve Generalization in Reinforcement Learning. 1427-1435 - Haruyuki Nakagawa, Yoshitaka Miyatani, Asako Kanezaki:
Linking Vision and Multi-Agent Communication through Visible Light Communication using Event Cameras. 1436-1444 - Linas Nasvytis, Kai Sandbrink, Jakob N. Foerster, Tim Franzmeyer, Christian Schröder de Witt:
Rethinking Out-of-Distribution Detection for Reinforcement Learning: Advancing Methods for Evaluation and Detection. 1445-1453 - Manisha Natarajan, Chunyue Xue, Sanne van Waveren, Karen M. Feigh, Matthew C. Gombolay:
Mixed-Initiative Human-Robot Teaming under Suboptimality with Online Bayesian Adaptation. 1454-1462 - Benjamin A. Newman, Christopher Jason Paxton, Kris Kitani, Henny Admoni:
Bootstrapping Linear Models for Fast Online Adaptation in Human-Agent Collaboration. 1463-1472 - Tong Niu, Weihao Zhang, Rong Zhao:
Solution-oriented Agent-based Models Generation with Verifier-assisted Iterative In-context Learning. 1473-1481 - Eura Nofshin, Siddharth Swaroop, Weiwei Pan, Susan A. Murphy, Finale Doshi-Velez:
Reinforcement Learning Interventions on Boundedly Rational Human Agents in Frictionful Tasks. 1482-1491 - Michael Oesterle, Tim Grams, Christian Bartelt, Heiner Stuckenschmidt:
RAISE the Bar: Restriction of Action Spaces for Improved Social Welfare and Equity in Traffic Management. 1492-1500 - Keisuke Okumura:
Engineering LaCAM*: Towards Real-time, Large-scale, and Near-optimal Multi-agent Pathfinding. 1501-1509 - Ninell Oldenburg, Tan Zhi-Xuan:
Learning and Sustaining Shared Normative Systems via Bayesian Rule Induction in Markov Games. 1510-1520 - Nicole Orzan, Erman Acar, Davide Grossi, Roxana Radulescu:
Emergent Cooperation under Uncertain Incentive Alignment. 1521-1530 - Nardine Osman, Mark d'Inverno:
A Computational Framework of Human Values. 1531-1539 - Soumyabrata Pal, Milind Tambe, Arun Sai Suggala, Karthikeyan Shanmugam, Aparna Taneja:
Improving Mobile Maternal and Child Health Care Programs: Collaborative Bandits for Time Slot Selection. 1540-1548 - Simone Parisi, Montaser Mohammedalamen, Alireza Kazemipour, Matthew E. Taylor, Michael Bowling:
Monitored Markov Decision Processes. 1549-1557 - Thomy Phan, Joseph Driscoll, Justin Romberg, Sven Koenig:
Confidence-Based Curriculum Learning for Multi-Agent Path Finding. 1558-1566 - Grzegorz Pierczynski, Stanislaw Szufa:
Single-Winner Voting with Alliances: Avoiding the Spoiler Effect. 1567-1575 - Subham Pokhriyal, Shweta Jain, Ganesh Ghalme, Swapnil Dhamal, Sujit Gujar:
Simultaneously Achieving Group Exposure Fairness and Within-Group Meritocracy in Stochastic Bandits. 1576-1584 - Antigoni Polychroniadou, Gabriele Cipriani, Richard Hua, Tucker Balch:
Atlas-X Equity Financing: Unlocking New Methods to Securely Obfuscate Axe Inventory Data Based on Differential Privacy. 1585-1592 - Nico Potyka, Yuqicheng Zhu, Yunjie He, Evgeny Kharlamov, Steffen Staab:
Robust Knowledge Extraction from Large Language Models using Social Choice Theory. 1593-1601 - Nicos Protopapas, Vahid Yazdanpanah, Enrico H. Gerding, Sebastian Stein:
Online Decentralised Mechanisms for Dynamic Ridesharing. 1602-1610 - Mathieu Reymond, Eugenio Bargiacchi, Diederik M. Roijers, Ann Nowé:
Interactively Learning the User's Utility for Best-Arm Identification in Multi-Objective Multi-Armed Bandits. 1611-1620 - Sebastian Rodriguez, John Thangarajah, Andrew Davey:
Design Patterns for Explainable Agents (XAg). 1621-1629 - Mikayel Samvelyan, Davide Paglieri, Minqi Jiang, Jack Parker-Holder, Tim Rocktäschel:
Multi-Agent Diagnostics for Robustness via Illuminated Diversity. 1630-1644 - Stefan Sarkadi, Peter R. Lewis:
The Triangles of Dishonesty: Modelling the Evolution of Lies, Bullshit, and Deception in Agent Societies. 1645-1653 - Nathaniel Sauerberg, Caspar Oesterheld:
Computing Optimal Commitments to Strategies and Outcome-Conditional Utility Transfers. 1654-1663 - Andreas Sauter, Nicolò Botteghi, Erman Acar, Aske Plaat:
CORE: Towards Scalable and Efficient Causal Discovery with Reinforcement Learning. 1664-1672 - Sangwon Seo, Vaibhav V. Unhelkar:
IDIL: Imitation Learning of Intent-Driven Expert Behavior. 1673-1682 - Marc Serramia, Natalia Criado, Michael Luck:
Multi-user Norm Consensus. 1683-1691 - Marc Serramia, Maite López-Sánchez, Juan A. Rodríguez-Aguilar, Stefano Moretti:
Value Alignment in Participatory Budgeting. 1692-1700 - Sanket Shah, Arun Sai Suggala, Milind Tambe, Aparna Taneja:
Efficient Public Health Intervention Planning Using Decomposition-Based Decision-focused Learning. 1701-1709 - Vade Shah, Jason R. Marden:
Battlefield Transfers in Coalitional Blotto Games. 1710-1717 - Aditya Shinde, Prashant Doshi:
Modeling Cognitive Biases in Decision-theoretic Planning for Active Cyber Deception. 1718-1726 - Shahaf S. Shperberg, Bo Liu, Peter Stone:
Relaxed Exploration Constrained Reinforcement Learning. 1727-1735 - Yash Shukla, Wenchang Gao, Vasanth Sarathy, Alvaro Velasquez, Robert Wright, Jivko Sinapov:
LgTS: Dynamic Task Sampling using LLM-generated Sub-Goals for Reinforcement Learning Agents. 1736-1744 - Arambam James Singh, Arvind Easwaran:
PAS: Probably Approximate Safety Verification of Reinforcement Learning Policy Using Scenario Optimization. 1745-1753 - Nikhil Kumar Singh, Indranil Saha:
Frugal Actor-Critic: Sample Efficient Off-Policy Deep Reinforcement Learning Using Unique Experiences. 1754-1762 - Tran Cao Son, Loc Pham, Enrico Pontelli:
On Dealing with False Beliefs and Maintaining KD45n Property. 1763-1771 - Yan Song, He Jiang, Haifeng Zhang, Zheng Tian, Weinan Zhang, Jun Wang:
Boosting Studies of Multi-Agent Reinforcement Learning on Google Research Football Environment: The Past, Present, and Future. 1772-1781 - Jean Springsteen, William Yeoh, Dino P. Christenson:
Algorithmic Filtering, Out-Group Stereotype, and Polarization on Social Media. 1782-1790 - Kefan Su, Siyuan Zhou, Jiechuan Jiang, Chuang Gan, Xiangjun Wang, Zongqing Lu:
Multi-Agent Alternate Q-Learning. 1791-1799 - Ankang Sun, Bo Li:
Allocating Contiguous Blocks of Indivisible Chores Fairly: Revisited. 1800-1808 - Hideaki Takahashi, Alex Fukunaga:
On the Transit Obfuscation Problem. 1809-1817 - Xinyu Tang, Hongtao Lv, Yingjie Gao, Fan Wu, Lei Liu, Lizhen Cui:
Towards Efficient Auction Design with ROI Constraints. 1818-1826 - Swapna Thorve, Henning S. Mortveit, Anil Vullikanti, Madhav V. Marathe, Samarth Swarup:
Assessing Fairness of Residential Dynamic Pricing for Electricity using Active Learning with Agent-based Simulation. 1827-1836 - Sz-Ting Tzeng, Nirav Ajmeri, Munindar P. Singh:
Norm Enforcement with a Soft Touch: Faster Emergence, Happier Agents. 1837-1846 - Filip Úradník, David Sychrovský, Jakub Cerný, Martin Cerný:
Reducing Optimism Bias in Incomplete Cooperative Games. 1847-1855 - Danai Vachtsevanou, Bruno de Lima, Andrei Ciortea, Jomi Fred Hübner, Simon Mayer, Jérémy Lemée:
Enabling BDI Agents to Reason on a Dynamic Action Repertoire in Hypermedia Environments. 1856-1864 - Aravind Venugopal, Stephanie Milani, Fei Fang, Balaraman Ravindran:
MABL: Bi-Level Latent-Variable World Model for Sample-Efficient Multi-Agent Reinforcement Learning. 1865-1873 - Chenmin Wang, Peng Li, Yulong Zeng, Xuepeng Fan:
Optimal Flash Loan Fee Function with Respect to Leverage Strategies. 1874-1882 - Ying Wang, Houyu Zhou, Minming Li:
Positive Intra-Group Externalities in Facility Location. 1883-1891 - Yongzhao Wang, Michael P. Wellman:
Generalized Response Objectives for Strategy Exploration in Empirical Game-Theoretic Analysis. 1892-1900 - Francis Rhys Ward, Matt MacDermott, Francesco Belardinelli, Francesca Toni, Tom Everitt:
The Reasons that Agents Act: Intention and Instrumental Goals. 1901-1909 - Jamison W. Weber, Dhanush R. Giriyan, Devendra R. Parkar, Dimitri P. Bertsekas, Andréa W. Richa:
Distributed Online Rollout for Multivehicle Routing in Unmapped Environments. 1910-1918 - Jannis Weil, Zhenghua Bao, Osama Abboud, Tobias Meuser:
Towards Generalizability of Multi-Agent Reinforcement Learning in Graphs with Recurrent Message Passing. 1919-1927 - Eric Wete, Joel Greenyer, Daniel Kudenko, Wolfgang Nejdl:
Multi-Robot Motion and Task Planning in Automotive Production Using Controller-based Safe Reinforcement Learning. 1928-1937 - Xiaoliang Wu, Qilong Feng, Ziyun Huang, Jinhui Xu, Jianxin Wang:
New Algorithms for Distributed Fair k-Center Clustering: Almost Accurate as Sequential Algorithms. 1938-1946 - Xiaoqiang Wu, Qingling Zhu, Qiuzhen Lin, Weineng Chen, Jianqiang Li:
Adaptive Evolutionary Reinforcement Learning Algorithm with Early Termination Strategy. 1947-1955 - Yaoxin Wu, Mingfeng Fan, Zhiguang Cao, Ruobin Gao, Yaqing Hou, Guillaume Sartoretti:
Collaborative Deep Reinforcement Learning for Solving Multi-Objective Vehicle Routing Problems. 1956-1965 - Mengwei Xu, Louise A. Dennis, Mustafa A. Mustafa:
Safeguard Privacy for Minimal Data Collection with Trustworthy Autonomous Agents. 1966-1974 - Yongxin Xu, Shangshang Wang, Hengquan Guo, Xin Liu, Ziyu Shao:
Learning to Schedule Online Tasks with Bandit Feedback. 1975-1983 - Zhaolin Xue, Lihua Zhang, Zhiyan Dong:
Successively Pruned Q-Learning: Using Self Q-function to Reduce the Overestimation. 1984-1992 - Yibin Yang, Mingfeng Fan, Chengyang He, Jianqiang Wang, Heye Huang, Guillaume Sartoretti:
Attention-based Priority Learning for Limited Time Multi-Agent Path Finding. 1993-2001 - Yucheng Yang, Tianyi Zhou, Lei Han, Meng Fang, Mykola Pechenizkiy:
Automatic Curriculum for Unsupervised Reinforcement Learning. 2002-2010 - Yunhao Yang, Cyrus Neary, Ufuk Topcu:
Multimodal Pretrained Models for Verifiable Sequential Decision-Making: Planning, Grounding, and Perception. 2011-2019 - Zewen Yang, Xiaobing Dai, Akshat Dubey, Sandra Hirche, Georges Hattab:
Whom to Trust? Elective Learning for Distributed Gaussian Process Regression. 2020-2028 - Zhaoxing Yang, Haiming Jin, Yao Tang, Guiyun Fan:
Risk-Aware Constrained Reinforcement Learning with Non-Stationary Policies. 2029-2037 - Batuhan Yardim, Artur Goldman, Niao He:
When is Mean-Field Reinforcement Learning Tractable and Relevant? 2038-2046 - Ahad N. Zehmakan, Xiaotian Zhou, Zhongzhi Zhang:
Viral Marketing in Social Networks with Competing Products. 2047-2056 - Ahad N. Zehmakan:
Majority-based Preference Diffusion on Social Networks. 2057-2065 - Chenyuan Zhang, Charles Kemp, Nir Lipovetzky:
Human Goal Recognition as Bayesian Inference: Investigating the Impact of Actions, Timing, and Goal Solvability. 2066-2074 - Mingyue Zhang, Nianyu Li, Jialong Li, Jiachun Liao, Jiamou Liu:
Memory-Based Resilient Control Against Non-cooperation in Multi-agent Flocking. 2075-2084 - Zhicheng Zhang, Yancheng Liang, Yi Wu, Fei Fang:
MESA: Cooperative Meta-Exploration in Multi-Agent Learning through Exploiting State-Action Space Structure. 2085-2093 - Tan Zhi-Xuan, Lance Ying, Vikash Mansinghka, Joshua B. Tenenbaum:
Pragmatic Instruction Following and Goal Assistance via Cooperative Language-Guided Inverse Planning. 2094-2103 - Zhiqiang Zhuang, Kewen Wang, Zhe Wang, Junhu Wang, Yinong Yang:
Maximising the Influence of Temporary Participants in Opinion Formation. 2104-2110
Extended Abstract
- Marwa Abdulhai, Micah Carroll, Justin Svegliato, Anca D. Dragan, Sergey Levine:
Defining Deception in Decision Making. 2111-2113 - Michael Akintunde, Vahid Yazdanpanah, Asieh Salehi Fathabadi, Corina Cîrstea, Mehdi Dastani, Luc Moreau:
Actual Trust in Multiagent Systems. 2114-2116 - Gianvincenzo Alfano, Sergio Greco, Francesco Parisi, Irina Trubitsyna:
On General Epistemic Abstract Argumentation Frameworks. 2117-2119 - Alessandro Aloisio, Vittorio Bilò, Antonio Mario Caruso, Michele Flammini, Cosimo Vinci:
Approximately Fair Allocation of Indivisible Items with Random Valuations. 2120-2122 - Philipp Altmann, Adelina Bärligea, Jonas Stein, Michael Kölle, Thomas Gabor, Thomy Phan, Claudia Linnhoff-Popien:
Quantum Circuit Design: A Reinforcement Learning Challenge. 2123-2125 - Ramsundar Anandanarayanan, Swaprava Nath, Rohit Vaish:
Charging Electric Vehicles Fairly and Efficiently. 2126-2128 - Ben Aoki-Sherwood, Catherine Bregou, David Liben-Nowell, Kiran Tomlinson, Thomas Zeng:
Bounding Consideration Probabilities in Consider-Then-Choose Ranking Models. 2129-2131 - Iosif Apostolakis, Zeynep G. Saribatur, Johannes Peter Wallner:
Abstracting Assumptions in Structured Argumentation. 2132-2134 - Ben Armstrong, Kate Larson:
Liquid Democracy for Low-Cost Ensemble Pruning. 2135-2137 - Viviana Arrigoni, Giulio Attenni, Novella Bartolini, Matteo Finelli, Gaia Maselli:
MiKe: Task Scheduling for UAV-based Parcel Delivery. 2138-2140 - Ayhan Alp Aydeniz, Enrico Marchesini, Christopher Amato, Kagan Tumer:
Entropy Seeking Constrained Multiagent Reinforcement Learning. 2141-2143 - Amirreza Bagheridelouee, Marzie Nilipour, Masoud Seddighin, Maziar Shamsipour:
Metric Distortion Under Public-Spirited Voting. 2144-2146 - Martina Baiardi, Samuele Burattini, Giovanni Ciatto, Danilo Pianini, Andrea Omicini, Alessandro Ricci:
Concurrency Model of BDI Programming Frameworks: Why Should We Control It? 2147-2149 - Ridhima Bector, Abhay Aradhya, Chai Quek, Zinovi Rabinovich:
Adaptive Discounting of Training Time Attacks. 2150-2152 - Márton Benedek, Péter Biró, Gergely Csáji, Matthew Johnson, Daniël Paulusma, Xin Ye:
Computing Balanced Solutions for Large International Kidney Exchange Schemes when Cycle Length is Unbounded. 2153-2155 - Nicolas Bessone, Payam Zahadat, Kasper Støy:
Decentralized Control of Distributed Manipulators: An Information Diffusion Approach. 2156-2158 - Abhijat Biswas, Badal Arun Pardhi, Caleb Chuck, Jarrett Holtz, Scott Niekum, Henny Admoni, Alessandro Allievi:
Gaze Supervision for Mitigating Causal Confusion in Driving Agents. 2159-2161 - Arpita Biswas, Yiduo Ke, Samir Khuller, Quanquan C. Liu:
Fair Allocation of Conflicting Courses under Additive Utilities. 2162-2164 - Megha Bose, Praveen Paruchuri, Akshat Kumar:
Factored MDP based Moving Target Defense with Dynamic Threat Modeling. 2165-2167 - Jérôme Botoko Ekila, Jens Nevens, Lara Verheyen, Katrien Beuls, Paul Van Eecke:
Decentralised Emergence of Robust and Adaptive Linguistic Conventions in Populations of Autonomous Agents Grounded in Continuous Worlds. 2168-2170 - Pedro Braga, Georgios Chionas, Piotr Krysta, Stefanos Leonardos, Georgios Piliouras, Carmine Ventre:
Who gets the Maximal Extractable Value? A Dynamic Sharing Blockchain Mechanism. 2171-2173 - Berk Buzcu, Emre Kuru, Reyhan Aydogan:
User-centric Explanation Strategies for Interactive Recommenders. 2174-2176 - Titas Chakraborty, Parth Shettiwar:
Non Stationary Bandits with Periodic Variation. 2177-2179 - Hau Chan, Xinliang Fu, Minming Li, Chenhao Wang:
Mechanism Design for Reducing Agent Distances to Prelocated Facilities. 2180-2182 - Shao-Hung Chan, Zhe Chen, Dian-Lun Lin, Yue Zhang, Daniel Harabor, Sven Koenig, Tsung-Wei Huang, Thomy Phan:
Anytime Multi-Agent Path Finding using Operation Parallelism in Large Neighborhood Search. 2183-2185 - Prabhat Kumar Chand, Apurba Das, Anisur Rahaman Molla:
Agent-Based Triangle Counting and Its Applications in Anonymous Graphs. 2186-2188 - Bin Chen, Zehong Cao:
HLG: Bridging Human Heuristic Knowledge and Deep Reinforcement Learning for Optimal Agent Performance. 2189-2191 - Jiehua Chen, William S. Zwicker:
Cutsets and EF1 Fair Division of Graphs. 2192-2194 - Siqi Chen, Jianing Zhao, Kai Zhao, Gerhard Weiss, Fengyun Zhang, Ran Su, Yang Dong, Daqian Li, Kaiyou Lei:
ANOTO: Improving Automated Negotiation via Offline-to-Online Reinforcement Learning. 2195-2197 - Yihong Chen, Cong Wang, Tianpei Yang, Meng Wang, Yingfeng Chen, Jifei Zhou, Chaoyi Zhao, Xinfeng Zhang, Zeng Zhao, Changjie Fan, Zhipeng Hu, Rong Xiong, Long Zeng:
Mastering Robot Control through Point-based Reinforcement Learning with Pre-training. 2198-2200 - Yuxin Chen, Chen Tang, Ran Tian, Chenran Li, Jinning Li, Masayoshi Tomizuka, Wei Zhan:
Quantifying Agent Interaction in Multi-agent Reinforcement Learning for Cost-efficient Generalization. 2201-2203 - Zixuan Chen, Ze Ji, Shuyang Liu, Jing Huo, Yiyu Chen, Yang Gao:
Cognizing and Imitating Robotic Skills via a Dual Cognition-Action Architecture. 2204-2206 - Janvi Chhabra, Jayati Deshmukh, Srinath Srinivasa:
Modelling the Dynamics of Subjective Identity in Allocation Games. 2207-2209 - Yu Quan Chong, Jiaoyang Li, Katia P. Sycara:
Optimal Task Assignment and Path Planning using Conflict-Based Search with Precedence and Temporal Constraints. 2210-2212 - Moumita Choudhury, Sandhya Saisubramanian, Hao Zhang, Shlomo Zilberstein:
Minimizing Negative Side Effects in Cooperative Multi-Agent Systems using Distributed Coordination. 2213-2215 - Alexandra Cimpean, Catholijn M. Jonker, Pieter Libin, Ann Nowé:
A Reinforcement Learning Framework for Studying Group and Individual Fairness. 2216-2218 - Saar Cohen, Noa Agmon:
Near-Optimal Online Resource Allocation in the Random-Order Model. 2219-2221 - Stephen Cranefield, Sriashalya Srivathsan, Jeremy Pitt:
Inferring Lewisian Common Knowledge using Theory of Mind Reasoning in a Forward-chaining Rule Engine. 2222-2224 - Sankarshan Damle, Sujit Gujar:
Analyzing Crowdfunding of Public Projects Under Dynamic Beliefs. 2225-2227 - Sankarshan Damle, Varul Srivastava, Sujit Gujar:
No Transaction Fees? No Problem! Achieving Fairness in Transaction Fee Mechanism Design. 2228-2230 - Gökçe Dayanikli, Mathieu Laurière, Jiacheng Zhang:
Deep Learning for Population-Dependent Controls in Mean Field Control Problems with Common Noise. 2231-2233 - Dave de Jonge, Laura Rodriguez Cima:
Attila: A Negotiating Agent for the Game of Diplomacy, Based on Purely Symbolic A.I. 2234-2236 - Pankaj Deoli, Rohit Kumar, Axel Vierling, Karsten Berns:
Evaluation of Robustness of Off-Road Autonomous Driving Segmentation against Adversarial Attacks: A Dataset-Centric Study. 2237-2239 - Ayse Mutlu Derya:
A Comparison of the Myerson Value and the Position Value. 2240-2242 - Mauricio Diaz-Ortiz Jr., Benjamin Kempinski, Daphne Cornelisse, Yoram Bachrach, Tal Kachman:
Pruning Neural Networks Using Cooperative Game Theory. 2243-2245 - Edith Elkind, Svetlana Obraztsova, Nicholas Teh:
Verifying Proportionality in Temporal Voting. 2246-2248 - Emre Erdogan, Rineke Verbrugge, Pinar Yolum:
Computational Theory of Mind with Abstractions for Effective Human-Agent Collaboration. 2249-2251 - Erwan Escudie, Laëtitia Matignon, Jacques Saraydaryan:
Attention Graph for Multi-Robot Social Navigation with Deep Reinforcement Learning. 2252-2254 - Piotr Faliszewski, Lukasz Janeczko, Andrzej Kaczmarczyk, Grzegorz Lisowski, Piotr Skowron, Stanislaw Szufa:
Strategic Cost Selection in Participatory Budgeting. 2255-2257 - Michael Y. Fatemi, Wesley A. Suttle, Brian M. Sadler:
Deceptive Path Planning via Reinforcement Learning with Graph Neural Networks. 2258-2260 - Andrew Festa, Gaurav Dixit, Kagan Tumer:
Influence-Focused Asymmetric Island Model. 2261-2263 - Tamara C. P. Florijn, Pinar Yolum, Tim Baarslag:
A Negotiator's Backup Plan: Optimal Concessions with a Reservation Value. 2264-2266 - Tim French:
Aleatoric Predicates: Reasoning about Marbles. 2267-2269 - Rustam Galimullin, Louwe B. Kuijer:
Synthesizing Social Laws with ATL Conditions. 2270-2272 - Zemuy Tesfay Gebrekidan, Sebastian Stein, Timothy J. Norman:
Combinatorial Client-Master Multiagent Deep Reinforcement Learning for Task Offloading in Mobile Edge Computing. 2273-2275 - Gaël Gendron, Yang Chen, Mitchell Rogers, Yiping Liu, Mihailo Azhar, Shahrokh Heidari, David Arturo Soriano Valdez, Kobe Knowles, Padriac O'Leary, Simon Eyre, Michael Witbrock, Gillian Dobbie, Jiamou Liu, Patrice Delmas:
Behaviour Modelling of Social Animals via Causal Structure Discovery and Graph Neural Networks. 2276-2278 - Minghong Geng, Shubham Pateria, Budhitama Subagdja, Ah-Hwee Tan:
Benchmarking MARL on Long Horizon Sequential Multi-Objective Tasks. 2279-2281 - Hafez Ghaemi, Hamed Kebriaei, Alireza Ramezani Moghaddam, Majid Nili Ahmadabadi:
Risk-Sensitive Multi-Agent Reinforcement Learning in Network Aggregative Markov Games. 2282-2284 - Zifan Gong, Minming Li, Houyu Zhou:
Facility Location Games with Task Allocation. 2285-2287 - Everardo Gonzalez, Siddarth Viswanathan, Kagan Tumer:
Indirect Credit Assignment in a Multiagent System. 2288-2290 - Alexander W. Goodall, Francesco Belardinelli:
Leveraging Approximate Model-based Shielding for Probabilistic Safety Guarantees in Continuous Environments. 2291-2293 - Alexey Gorbatovski, Sergey V. Kovalchuk:
Reinforcement Learning for Question Answering in Programming Domain using Public Community Scoring as a Human Feedback. 2294-2296 - Youssef Hamadi, Gauthier Picard:
Towards Socially-Acceptable Multi-Criteria Resolution of the 4D-Contracts Repair Problem. 2297-2299 - Matthew P. Hare, Doug Salt, Ric Colasanti, Richard Milton, Mike Batty, Alison J. Heppenstall, Gary Polhill:
Taking Agent-Based Social Simulation to the Next Level Using Exascale Computing: Potential Use-Cases, Capacity Requirements and Threats. 2300-2302 - Somnath Hazra, Pallab Dasgupta, Soumyajit Dey:
Addressing Permutation Challenges in Multi-Agent Reinforcement Learning. 2303-2305 - Hadi Hosseini, Joshua Kavner, Tomasz Was, Lirong Xia:
Distribution of Chores with Information Asymmetry. 2306-2308 - Mohammad T. Irfan, Hau Chan, Jared Soundy:
Computing Nash Equilibria in Multidimensional Congestion Games. 2309-2311 - Kazi Ashik Islam, Da Qi Chen, Madhav V. Marathe, Henning S. Mortveit, Samarth Swarup, Anil Vullikanti:
Strategic Routing and Scheduling for Evacuations. 2312-2314 - Weibo Jiang, Shaohui Li, Zhi Li, Yuxin Ke, Zhizhuo Jiang, Yaowen Li, Yu Liu:
Dual-Policy-Guided Offline Reinforcement Learning with Optimal Stopping. 2315-2317 - Joanna Kaczmarek, Jörg Rothe:
NPPP-Completeness of Control by Adding Players to Change the Penrose-Banzhaf Power Index in Weighted Voting Games. 2318-2320 - Qitong Kang, Fuyong Wang, Zhongxin Liu, Zengqiang Chen:
TIMAT: Temporal Information Multi-Agent Transformer. 2321-2323 - Bruce M. Kapron, Koosha Samieefar:
On the Computational Complexity of Quasi-Variational Inequalities and Multi-Leader-Follower Games. 2324-2326 - Yasushi Kawase, Bodhayan Roy, Mohammad Azharuddin Sanpui:
Contiguous Allocation of Binary Valued Indivisible Items on a Path. 2327-2329 - Saad Khan, Mayank Baranwal, Srikant Sukumar:
Decentralized Safe Control for Multi-Robot Navigation in Dynamic Environments with Limited Sensing. 2330-2332 - Chaitanya Kharyal, Sai Krishna Gottipati, Tanmay Kumar Sinha, Srijita Das, Matthew E. Taylor:
GLIDE-RL: Grounded Language Instruction through DEmonstration in RL. 2333-2335 - Daisuke Kikuta, Hiroki Ikeuchi, Kengo Tajiri, Yuta Toyama, Masaki Nakamura, Yuusuke Nakano:
Electric Vehicle Routing for Emergency Power Supply with Deep Reinforcement Learning. 2336-2338 - Akshat Kumar:
Difference of Convex Functions Programming for Policy Optimization in Reinforcement Learning. 2339-2341 - Pankaj Kumar:
Deep Hawkes Process for High-Frequency Market Making. 2342-2344 - Yatharth Kumar, Sarfaraz Equbal, Rohit Gurjar, Swaprava Nath, Rohit Vaish:
Fair Scheduling of Indivisible Chores. 2345-2347 - Igor Kuznetsov:
Guided Exploration in Reinforcement Learning via Monte Carlo Critic Optimization. 2348-2350 - Jean-Marie Lagniez, Emmanuel Lonca, Jean-Guy Mailly:
A SAT-based Approach for Argumentation Dynamics. 2351-2353 - Daji Landis, Nikolaj I. Schwartzbach:
Which Games are Unaffected by Absolute Commitments? 2354-2356 - Shiqi Lei, Kanghoon Lee, Linjing Li, Jinkyoo Park, Jiachen Li:
ELA: Exploited Level Augmentation for Offline Learning in Zero-Sum Games. 2357-2359 - Dapeng Li, Zhiwei Xu, Bin Zhang, Guangchong Zhou, Zeren Zhang, Guoliang Fan:
From Explicit Communication to Tacit Cooperation: A Novel Paradigm for Cooperative MARL. 2360-2362 - Yansong Li, Shuo Han:
Efficient Collaboration with Unknown Agents: Ignoring Similar Agents without Checking Similarity. 2363-2365 - Ruixi Luo, Kai Jin, Zelin Ye:
Simple k-crashing Plan with a Good Approximation Ratio. 2366-2368 - Jiafei Lyu, Le Wan, Xiu Li, Zongqing Lu:
Towards Understanding How to Reduce Generalization Gap in Visual Reinforcement Learning. 2369-2371 - Jayakrishnan Madathil, Neeldhara Misra, Yash More:
Opinion Diffusion on Society Graphs Based on Approval Ballots. 2372-2374 - Yi Mao, Andrew Perrault:
Time-Constrained Restless Multi-Armed Bandits with Applications to City Service Scheduling. 2375-2377 - Mathieu Mari, Michal Pawlowski, Runtian Ren, Piotr Sankowski:
Multi-level Aggregation with Delays and Stochastic Arrivals. 2378-2380 - Yongsheng Mei, Hanhan Zhou, Tian Lan:
Projection-Optimal Monotonic Value Function Factorization in Multi-Agent Reinforcement Learning. 2381-2383 - Daniel Melcer, Christopher Amato, Stavros Tripakis:
Shield Decentralization for Safe Reinforcement Learning in General Partially Observable Multi-Agent Environments. 2384-2386 - Alexander Mendelsohn, Donald Sofge, Michael W. Otte:
Enhancing Search and Rescue Capabilities in Hazardous Communication-Denied Environments through Path-Based Sensors with Backtracking. 2387-2389 - Karl Jochen Micheel, Anaëlle Wilczynski:
Fairness in Repeated House Allocation. 2390-2392 - David Milec, Ondrej Kubícek, Viliam Lisý:
Continual Depth-limited Responses for Computing Counter-strategies in Sequential Games. 2393-2395 - Glareh Mir, Michael Beetz:
Simulated Robotic Soft Body Manipulation. 2396-2398 - Calarina Muslimani, Matthew E. Taylor:
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning. 2399-2401 - Yu Niu, Hengxu Zhao, Lei Yu:
MA-MIX: Value Function Decomposition for Cooperative Multiagent Reinforcement Learning Based on Multi-Head Attention Mechanism. 2402-2404 - Alberto Olivares Alarcos, Sergi Foix, Júlia Borràs Sol, Gerard Canal, Guillem Alenyà:
Ontological Modeling and Reasoning for Comparison and Contrastive Narration of Robot Plans. 2405-2407 - Daniele Orner, Elizabeth Akinyi Ondula, Nick Mumero Mwangi, Richa Goyal:
Sentimental Agents: Combining Sentiment Analysis and Non-Bayesian Updating for Cooperative Decision-Making. 2408-2410 - Pranavi Pathakota, Hardik Meisheri, Harshad Khadilkar:
DCT: Dual Channel Training of Action Embeddings for Reinforcement Learning with Large Discrete Action Spaces. 2411-2413 - Lukasz Pelcner, Matheus Aparecido do Carmo Alves, Leandro Soriano Marcolino, Paula A. Harrison, Peter M. Atkinson:
Incentive-based MARL Approach for Commons Dilemmas in Property-based Environments. 2414-2416 - Ava Pettet, Yunuo Zhang, Baiting Luo, Kyle Hollins Wray, Hendrik Baier, Aron Laszka, Abhishek Dubey, Ayan Mukhopadhyay:
Decision Making in Non-Stationary Environments with Policy-Augmented Search. 2417-2419 - Xuan Kien Phung, Sylvie Hamel:
Optimal Majority Rules and Quantitative Condorcet Properties of Setwise Kemeny Voting Schemes. 2420-2422 - Rafael Pina, Varuna De Silva, Corentin Artaud, Xiaolan Liu:
Fully Independent Communication in Multi-Agent Reinforcement Learning. 2423-2425 - Ram Rachum, Yonatan Nakar, Bill Tomlinson, Nitay Alon, Reuth Mirsky:
Emergent Dominance Hierarchies in Reinforcement Learning Agents. 2426-2428 - Ashish Rana, Michael Oesterle, Jannik Brinkmann:
GOV-REK: Governed Reward Engineering Kernels for Designing Robust Multi-Agent Reinforcement Learning Systems. 2429-2431 - John Randolph, Amy Greenwald, Denizalp Goktas:
Banzhaf Power in Hierarchical Games. 2432-2434 - Maxime Reynouard, Olga Gorelkina, Rida Laraki:
BAR Nash Equilibrium and Application to Blockchain Design. 2435-2437 - Erin E. Richardson, Savannah Lynn Buchner, Jacob R. Kintz, Torin K. Clark, Allison P. Anderson:
Psychophysiological Models of Cognitive States Can Be Operator-Agnostic. 2438-2440 - Stefan Roesch, Stefanos Leonardos, Yali Du:
The Selfishness Level of Social Dilemmas. 2441-2443 - Alexander Rutherford, Benjamin Ellis, Matteo Gallici, Jonathan Cook, Andrei Lupu, Garðar Ingvarsson, Timon Willi, Akbir Khan, Christian Schröder de Witt, Alexandra Souly, Saptarashmi Bandyopadhyay, Mikayel Samvelyan, Minqi Jiang, Robert T. Lange, Shimon Whiteson, Bruno Lacerda, Nick Hawes, Tim Rocktäschel, Chris Lu, Jakob N. Foerster:
JaxMARL: Multi-Agent RL Environments and Algorithms in JAX. 2444-2446 - Yael Sabato, Amos Azaria, Noam Hazon:
Source Detection in Networks using the Stationary Distribution of a Markov Chain. 2447-2449 - Karthik Sama, Jayati Deshmukh, Srinath Srinivasa:
Social Identities and Responsible Agency. 2450-2452 - Pedro P. Santos, Diogo S. Carvalho, Miguel Vasco, Alberto Sardinha, Pedro A. Santos, Ana Paiva, Francisco S. Melo:
Centralized Training with Hybrid Execution in Multi-Agent Reinforcement Learning. 2453-2455 - Anindya Sarkar, Alex DiChristofano, Sanmay Das, Patrick J. Fowler, Nathan Jacobs, Yevgeniy Vorobeychik:
Geospatial Active Search for Preventing Evictions. 2456-2458 - Junning Shao, Siwei Wang, Zhixuan Fang:
Balanced and Incentivized Learning with Limited Shared Information in Multi-agent Multi-armed Bandit. 2459-2461 - Matthew Sheldon, Dario Paccagnan, Giuliano Casale:
Cournot Queueing Games with Applications to Mobility Systems. 2462-2464 - Haochen Shi, Zhiyuan Sun, Xingdi Yuan, Marc-Alexandre Côté, Bang Liu:
OPEx: A Large Language Model-Powered Framework for Embodied Instruction Following. 2465-2467 - Jacobus Smit, Fernando P. Santos:
Fairness and Cooperation between Independent Reinforcement Learners through Indirect Reciprocity. 2468-2470 - Sambhav Solanki, Sujit Gujar, Shweta Jain:
Fairness and Privacy Guarantees in Federated Contextual Bandits. 2471-2473 - Archit Sood, Shweta Jain, Sujit Gujar:
Fairness of Exposure in Online Restless Multi-armed Bandits. 2474-2476 - Timo Speith:
Unlocking the Potential of Machine Ethics with Explainability. 2477-2479 - Gogulapati Sreedurga:
Hybrid Participatory Budgeting: Divisible, Indivisible, and Beyond. 2480-2482 - Varul Srivastava, Sujit Gujar:
Decent-BRM: Decentralization through Block Reward Mechanisms. 2483-2485 - Mihail Stojanovski, Nadjet Bourdache, Grégory Bonnet, Abdel-Illah Mouaddib:
Ethical Markov Decision Processes with Moral Worth as Rewards. 2486-2488 - Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder, Tuomas Sandholm:
A Multiagent Path Search Algorithm for Large-Scale Coalition Structure Generation. 2489-2491 - Redha Taguelmimt, Samir Aknine, Djamila Boukredera, Narayan Changder, Tuomas Sandholm:
Efficient Size-based Hybrid Algorithm for Optimal Coalition Structure Generation. 2492-2494 - Kenjiro Takazawa:
Pure Nash Equilibria in Weighted Congestion Games with Complementarities and Beyond. 2495-2497 - Huijie Tang, Federico Berto, Zihan Ma, Chuanbo Hua, Kyuree Ahn, Jinkyoo Park:
HiMAP: Learning Heuristics-Informed Policies for Large-Scale Multi-Agent Pathfinding. 2498-2500 - Peng Tang, Lifan Wang, Weidong Qiu, Zheng Huang, Qiangmin Wang:
Fuzzy Clustered Federated Learning Under Mixed Data Distributions. 2501-2503 - Michael J. Tarlton, Gustavo B. M. Mello, Anis Yazidi:
Neurological Based Timing Mechanism for Reinforcement Learning. 2504-2506 - Jayden Teoh Jing Teoh, Wenjun Li, Pradeep Varakantham:
Unifying Regret and State-Action Space Coverage for Effective Unsupervised Environment Design. 2507-2509 - Kazunori Terada, Yasuo Noma, Masanori Hattori:
Persuasion by Shaping Beliefs about Multidimensional Features of a Thing. 2510-2512 - Emanuel Tewolde, Vincent Conitzer:
Game Transformations That Preserve Nash Equilibria or Best-Response Sets. 2513-2515 - Sheng Tian, Hong Shen, Yuan Tian, Hui Tian:
Consensus of Nonlinear Multi-Agent Systems with Semi-Markov Switching Under DoS Attacks. 2516-2518 - Jinyun Tong, Bart de Keijzer, Carmine Ventre:
Reducing Systemic Risk in Financial Networks through Donations. 2519-2521 - Maxime Toquebiau, Nicolas Bredèche, Faïz Ben Amar, Jae-Yun Jun:
Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent Deep Reinforcement Learning. 2522-2524 - Sharlin Utke, Jeremie Houssineau, Giovanni Montana:
Embracing Relational Reasoning in Multi-Agent Actor-Critic. 2525-2527 - Pascal R. van der Vaart, Neil Yorke-Smith, Matthijs T. J. Spaan:
Bayesian Ensembles for Exploration in Deep Q-Learning. 2528-2530 - Jarod Vanderlynden, Philippe Mathieu, Romain Warlop:
Understanding the Impact of Promotions on Consumer Behavior. 2531-2533 - Maya Viswanathan, Ruta Mehta:
On the existence of EFX under picky or non-differentiative agents. 2534-2536 - Khaing Phyo Wai, Minghong Geng, Shubham Pateria, Budhitama Subagdja, Ah-Hwee Tan:
Explaining Sequences of Actions in Multi-agent Deep Reinforcement Learning Models. 2537-2539 - Thayne T. Walker, Nathan R. Sturtevant, Ariel Felner:
Clique Analysis and Bypassing in Continuous-Time Conflict-Based Search. 2540-2542 - Chen Wang, Sarah M. Erfani, Tansu Alpcan, Christopher Leckie:
Detecting Anomalous Agent Decision Sequences Based on Offline Imitation Learning. 2543-2545 - Chenxu Wang, Zilong Chen, Huaping Liu:
On the Utility of External Agent Intention Predictor for Human-AI Coordination. 2546-2548 - Wenlong Wang, Thomas Pfeiffer:
Decision Market Based Learning for Multi-agent Contextual Bandit Problems. 2549-2551 - Xinrun Wang, Chang Yang, Shuxin Li, Pengdeng Li, Xiao Huang, Hau Chan, Bo An:
Reinforcement Nash Equilibrium Solver. 2552-2554 - Sam Williams, Jyotirmoy Deshmukh:
Potential Games on Cubic Splines for Multi-Agent Motion Planning of Autonomous Agents. 2555-2557 - Binghan Wu, Wei Bao, Bing Zhou:
Competitive Analysis of Online Facility Open Problem. 2558-2560 - Zida Wu, Mathieu Laurière, Samuel Jia Cong Chua, Matthieu Geist, Olivier Pietquin, Ankur Mehta:
Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning. 2561-2563 - Tianyi Yang, Yuxiang Zhai, Dengji Zhao, Xinwei Song, Miao Li:
Truthful and Stable One-sided Matching on Networks. 2564-2566 - Yongjie Yang:
On the Complexity of Candidates-Embedded Multiwinner Voting under the Hausdorff Function. 2567-2569 - Wentao Ye, Bo Liu, Yuan Luo, Jianwei Huang:
Dual Role AoI-based Incentive Mechanism for HD map Crowdsourcing. 2570-2572 - Jhih-Ching Yeh, Von-Wun Soo:
Toward Socially Friendly Autonomous Driving Using Multi-agent Deep Reinforcement Learning. 2573-2575 - Hao Yin, Fan Chen, Hongjie He:
Solving Offline 3D Bin Packing Problem with Large-sized Bin via Two-stage Deep Reinforcement Learning. 2576-2578 - William Yue, Bo Liu, Peter Stone:
Overview of t-DGR: A Trajectory-Based Deep Generative Replay Method for Continual Learning in Decision Making. 2579-2581 - Haipeng Zhang, Zhiwen Wang, Na Li:
MATLight: Traffic Signal Coordinated Control Algorithm based on Heterogeneous-Agent Mirror Learning with Transformer. 2582-2584 - Hao Zhang, Tianpei Yang, Yan Zheng, Jianye Hao, Matthew E. Taylor:
PADDLE: Logic Program Guided Policy Reuse in Deep Reinforcement Learning. 2585-2587 - Huihui Zhang:
Bellman Momentum on Deep Reinforcement Learning. 2588-2590 - Kaifeng Zhang, Rui Zhao, Ziming Zhang, Yang Gao:
Auto-Encoding Adversarial Imitation Learning. 2591-2593 - Shiyao Zhang, Yuji Dong, Yichuan Zhang, Terry R. Payne, Jie Zhang:
Large Language Model Assissted Multi-Agent Dialogue for Ontology Alignment. 2594-2596 - Xianjie Zhang, Jiahao Sun, Chen Gong, Kai Wang, Yifei Cao, Hao Chen, Yu Liu:
Mutual Information as Intrinsic Reward of Reinforcement Learning Agents for On-demand Ride Pooling. 2597-2599 - Yao Zhang, Shanshan Zheng, Dengji Zhao:
Optimal Diffusion Auctions. 2600-2602 - Yirui Zhang, Zhixuan Fang:
Decentralized Competing Bandits in Many-to-One Matching Markets. 2603-2605 - Binyu Zhao, Wei Zhang, Zhaonian Zou:
Distance-Aware Attentive Framework for Multi-Agent Collaborative Perception in Presence of Pose Error. 2606-2608 - Kai Zhao, Jianye Hao, Yi Ma, Jinyi Liu, Yan Zheng, Zhaopeng Meng:
ENOTO: Improving Offline-to-Online Reinforcement Learning with Q-Ensembles. 2609-2611 - Xin Zhao, Jiaxin Li, Zhiwei Fang, Yuchen Guo, Jinyuan Zhao, Jie He, Wenlong Chen, Changping Peng, Guiguang Ding:
JDRec: Practical Actor-Critic Framework for Online Combinatorial Recommender System. 2612-2614 - Yangyang Zhao, Mehdi Dastani, Shihan Wang:
Bootstrapped Policy Learning: Goal Shaping for Efficient Task-oriented Dialogue Policy Learning. 2615-2617 - Yunfan Zhao, Nikhil Behari, Edward Hughes, Edwin Zhang, Dheeraj Nagaraj, Karl Tuyls, Aparna Taneja, Milind Tambe:
Towards Zero Shot Learning in Restless Multi-armed Bandits. 2618-2620 - Yiwen Zhu, Jinyi Liu, Wenya Wei, Qianyi Fu, Yujing Hu, Zhou Fang, Bo An, Jianye Hao, Tangjie Lv, Changjie Fan:
vMFER: von Mises-Fisher Experience Resampling Based on Uncertainty of Gradient Directions for Policy Improvement of Actor-Critic Algorithms. 2621-2623
AAAI Track
- Shiri Alouf-Heffetz, Tanmay Inamdar, Pallavi Jain, Nimrod Talmon, Yash More Hiren:
Controlling Delegations in Liquid Democracy. 2624-2632 - Roman Belaire, Pradeep Varakantham, Thanh Hong Nguyen, David Lo:
Regret-based Defense in Adversarial Reinforcement Learning. 2633-2640 - Zheng Chen, Bo Li, Minming Li, Guochuan Zhang:
Fair and Efficient Division of a Discrete Cake with Switching Utility Loss. 2641-2649 - Sumanta Dey, Briti Gangopadhyay, Pallab Dasgupta, Soumyajit Dey:
MAGNets: Micro-Architectured Group Neural Networks. 2650-2658 - Sushmita Gupta, Pallavi Jain, A. Mohanapriya, Vikash Tripathi:
Budget-feasible Egalitarian Allocation of Conflicting Jobs. 2659-2667
Blue Sky Ideas Track
- Tim Baarslag:
Multi-deal Negotiation. 2668-2673 - Jean-Loup Farges, Filipo Perotto, Gauthier Picard, Cédric Pralet, Cyrille de Lussy, Jonathan Guerra, Philippe Pavero, Fabrice Planchou:
Going Beyond Mono-Mission Earth Observation: Using the Multi-Agent Paradigm to Federate Multiple Missions. 2674-2678 - Ramon Fraga Pereira, Felipe Meneguzzi:
Empowering BDI Agents with Generalised Decision-Making. 2679-2683 - Behrad Koohy, Jan Buermann, Vahid Yazdanpanah, Pamela Briggs, Paul Pschierer-Barnfather, Enrico H. Gerding, Sebastian Stein:
Adaptive Incentive Engineering in Citizen-Centric AI. 2684-2689 - Emiliano Lorini:
Designing Artificial Reasoners for Communication. 2690-2695 - Rui Prada, Astrid C. Homan, Gerben A. van Kleef:
Towards Sustainable Human-Agent Teams: A Framework for Understanding Human-Agent Team Dynamics. 2696-2700 - Manon Revel, Niclas Boehmer, Rachael Colley, Markus Brill, Piotr Faliszewski, Edith Elkind:
Selecting Representative Bodies: An Axiomatic View. 2701-2705 - Alessandro Ricci, Stefano Mariani, Franco Zambonelli, Samuele Burattini, Cristiano Castelfranchi:
The Cognitive Hourglass: Agent Abstractions in the Large Models Era. 2706-2711 - Sebastian Rodriguez, John Thangarajah:
Explainable Agents (XAg) by Design. 2712-2716 - Peter Vamplew, Cameron Foale, Conor F. Hayes, Patrick Mannion, Enda Howley, Richard Dazeley, Scott Johnson, Johan Källström, Gabriel de Oliveira Ramos, Roxana Radulescu, Willem Röpke, Diederik M. Roijers:
Utility-Based Reinforcement Learning: Unifying Single-objective and Multi-objective Reinforcement Learning. 2717-2721
Doctoral Consortium
- Iosif Apostolakis:
Abstraction in Non-Monotonic Reasoning. 2722-2724 - Jérôme Botoko Ekila:
Emergence of Linguistic Conventions In Multi-Agent Systems Through Situated Communicative Interactions. 2725-2727 - Jiaxun Cui:
Communication and Generalization in Multi-Agent Learning. 2728-2730 - Yihan Dong:
The Multi-agent System based on LLM for Online Discussions. 2731-2733 - Tamara C. P. Florijn:
Negotiation Strategies for Combining Partials Deals in One-To-Many Negotiations. 2734-2736 - Minghong Geng:
Scaling up Cooperative Multi-agent Reinforcement Learning Systems. 2737-2739 - Victor Gimenez-Abalos:
Toward Explainable Agent Behaviour. 2740-2742 - Shivam Goel:
Towards building Autonomous AI Agents and Robots for Open World Environments. 2743-2745 - Bram Grooten:
Large Learning Agents: Towards Continually Aligned Robots with Scale in RL. 2746-2748 - Himanshu Gupta:
Efficient Continuous Space BeliefMDP Solutions for Navigation and Active Sensing. 2749-2751 - Balint Gyevnar:
Building Trustworthy Human-Centric Autonomous Systems Via Explanations. 2752-2754 - Baiting Luo:
Adaptive Decision-Making in Non-Stationary Markov Decision Processes. 2755-2757 - Yiwei Lyu:
Interactive Control and Decision-Making for Multi-Robots Systems. 2758-2760 - Eura Nofshin:
Leveraging Interpretable Human Models to Personalize AI Interventions for Behavior Change. 2761-2763 - Erin E. Richardson:
Predicting and Protecting the Cognitive Health of Operators in Isolated, Confined, and Extreme Environments. 2764-2766 - Pedro P. Santos:
Generalizing Objective-Specification in Markov Decision Processes. 2767-2769 - Jarrod Shipton:
Cooperative Multi-Agent Reinforcement Learning in Convention Reliant Environments. 2773-2775 - Yash Shukla:
Formal and Natural Language assisted Curriculum Generation for Reinforcement Learning Agents. 2776-2778 - Nicholas Teh:
Distributive and Temporal Fairness in Algorithmic Collective Decision-Making. 2779-2781 - Pascal R. van der Vaart:
Bayesian Model-Free Deep Reinforcement Learning. 2782-2784 - Gautham Vasan:
Autonomous Skill Acquisition for Robots Using Graduated Learning. 2785-2787 - Shiji Xing:
Allocating Resources with Imperfect Information. 2788-2790 - Zhicheng Zhang:
Advancing Sample Efficiency and Explainability in Multi-Agent Reinforcement Learning. 2791-2793
Demonstration Track
- Alexandry Augustin, Elnaz Shafipour, Sebastian Stein:
EVtonomy: A Personalised Route Planner for Electric Vehicles. 2794-2796 - Maxence de Rochechouart, Raed Abu Zitar, Amal El Fallah Seghrouchni, Frédéric Barbaresco:
End to End Camera only Drone Detection and Tracking Demo within a Multi-agent Framework with a CNN-LSTM Model for Range Estimation. 2797-2799 - Nathan Gavenski, Michael Luck, Odinaldo Rodrigues:
Imitation Learning Datasets: A Toolkit For Creating Datasets, Training Agents and Benchmarking. 2800-2802 - Moritz Graf, Thorsten Engesser, Bernhard Nebel:
A Symbolic Sequential Equilibria Solver for Game Theory Explorer. 2803-2805 - Sunghoon Hong, Deunsol Yoon, Whiyoung Jung, Jinsang Lee, Hyundam Yoo, Jiwon Ham, Suhyun Jung, Chanwoo Moon, Yeontae Jung, Kanghoon Lee, Woohyung Lim, Somin Jeon, Myounggu Lee, Sohui Hong, Jaesang Lee, Hangyoul Jang, Changhyun Kwak, Jeonghyeon Park, Changhoon Kang, Jungki Kim:
Naphtha Cracking Center Scheduling Optimization using Multi-Agent Reinforcement Learning. 2806-2808 - William Hunt, Toby Godfrey, Mohammad Divband Soorati:
Conversational Language Models for Human-in-the-Loop Multi-Robot Coordination. 2809-2811 - Mateusz Kaminski, Damian Kurpiewski, Wojciech Jamroga:
STV+KH: Towards Practical Verification of Strategic Ability for Knowledge and Information Flow. 2812-2814 - Artur Niewiadomski, Maciej Nazarczuk, Mateusz Przychodzki, Magdalena Kacprzak, Wojciech Penczek, Andrzej Zbrzezny:
SMT4SMTL: A Tool for SMT-Based Satisfiability Checking of SMTL. 2815-2817 - Yang Qiu, Ping Chen, Huiguo Zhang, Bo Huang, Di Wang, Zhiqi Shen:
Engaging the Elderly in Exercise with Agents: A Gamified Stationary Bike System for Sarcopenia Management. 2818-2820 - Adrián Tormos, Victor Gimenez-Abalos, Javier Vázquez-Salceda, Sergio Álvarez-Napagao:
pgeon applied to Overcooked-AI to explain agents' behaviour. 2821-2823
JAAMAS Track
- Cleber Jorge Amaral, Jomi F. Hübner, Stephen Cranefield:
Generating and Choosing Organizations for Multi-Agent Systems. 2824-2826 - Cyrille Berger, Patrick Doherty, Piotr Rudol, Mariusz Wzorek:
A Summary of the RGS⊕: an RDF Graph Synchronization System for Collaborative Robotics. 2827-2829 - Le Cong Dinh, David Henry Mguni, Long Tran-Thanh, Jun Wang, Yaodong Yang:
A Summary of Online Markov Decision Processes with Non-oblivious Strategic Adversary. 2830-2832 - Bin Li, Dong Hao, Dengji Zhao:
Extended Abstract of Diffusion Auction Design with Transaction Costs. 2833-2835 - Geeta Mahala, Özgür Kafali, Hoa Khanh Dam, Aditya Ghose, Munindar P. Singh:
Toward a Normative Approach for Resilient Multiagent Systems: A Summary. 2836-2838 - Nieves Montes, Michael Luck, Nardine Osman, Odinaldo Rodrigues, Carles Sierra:
Combining Theory of Mind and Abductive Reasoning in Agent-Oriented Programming. 2839-2841 - Jianglin Qiao, Dave de Jonge, Dongmo Zhang, Simeon Simoff, Carles Sierra, Bo Du:
Extended Abstract: Price of Anarchy of Traffic Assignment with Exponential Cost Functions. 2842-2844 - Changxi Zhu, Mehdi Dastani, Shihan Wang:
A Survey of Multi-Agent Deep Reinforcement Learning with Communication. 2845-2847
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.