Kale-ab Tessera

PhD Candidate, University of Edinburgh.

prof_pic.jpg

Seeking research internships (Summer/Fall 2026) in LLM agents, multi-agent systems, open-endedness, cooperative AI, and reinforcement learning. kaleabtessera@gmail.com · resumé

I am a third-year PhD candidate at the University of Edinburgh studying how foundation-model agents reason, coordinate, and fail in dynamic, open-ended multi-agent environments. My work sits at the intersection of reinforcement learning, multi-agent systems, and agentic models, with a focus on long-horizon interaction, coordination, robustness, and failure modes.

More broadly, I study a core question: as agents are deployed in increasingly open-ended settings, when and why does cooperation break down, and how can we make it robust?

I am advised by Amos Storkey, Tim Rocktäschel (UCL), and Aris Filos-Ratsikas, and I am affiliated with MARBLE (Multi-Agent, Reinforcement, Behaviour and Learning), where I co-organise the 🤖 RL & Agents Reading Group.

Before starting my PhD, I spent 2.5 years as a Research Engineer on the MARL team at InstaDeep, alongside broader experience in machine learning and software engineering. My background combines research on multi-agent reinforcement learning with current work on LLM agents in open-ended settings.

Recent work includes HyperMARL, accepted at NeurIPS 2025, which studies adaptive cooperation in MARL, and Probing Dec-POMDP Reasoning, accepted as an Oral at AAMAS 2026, which develops information-theoretic tools for testing whether multi-agent benchmarks require genuine decentralised reasoning.

Research Interests:

  • LLM agents: evaluating coordination, reasoning, and robustness in open-ended multi-agent settings.
  • Reinforcement learning for agents: training and evaluating adaptive policies for long-horizon interaction, tool use, and open-ended environments.
  • Multi-agent learning and cooperation: understanding when tasks require genuine decentralised reasoning and when coordination breaks down.

For more information, see my resumé and Google Scholar.

news

May, 2026 🌟 Presenting Probing Dec-POMDP Reasoning in Cooperative MARL at AAMAS 2026 in Paphos, Cyprus.
Dec, 2025 🌟 Presented HyperMARL: Adaptive Hypernetworks for Multi-Agent RL at NeurIPS 2025 in San Diego, US.
Sep, 2025 🗣️ Talk on “Algorithms and Benchmarks for Robust Multi-Agent Coordination” at the RAIL Lab, University of the Witwatersrand.
Aug, 2025 🏅 Remembering the Markov Property in Cooperative MARL won best poster (1st place) out of 278 submissions at the Deep Learning Indaba in Kigali, Rwanda.
Aug, 2025 📅 Co-Programme Chair for the Deep Learning Indaba and Head of Practicals and Tutorials in Kigali, Rwanda.
Aug, 2025 Our reading group is back – 🤖 RL & Agents Reading Group.
Aug, 2025 🌟 Presented Remembering the Markov Property in Cooperative MARL and HyperMARL: Adaptive Hypernetworks for Multi-Agent RL at RLC workshops in Edmonton, Canada.
Mar, 2025 🌟 Attended UK Multi-Agent Systems Symposium 2025 at King’s College London.
Aug, 2024 🗣️ Taught “Introduction to ML” at DLI.
Jul, 2024 🏅 Awarded a scholarship to attend the CIFAR Deep Learning and Reinforcement Learning (DLRL) Summer School in Toronto, Canada.
Jan, 2024 🗣️ Begin co-hosting the UOE RL reading group, YouTube.
Sep, 2023 🎓 Started my PhD at the University of Edinburgh (UOE), through the Informatics Global PhD Scholarship.
Aug, 2023 🛠️ PC member and Practicals Chair of DLI - notebooks 2023, RL Prac.
May, 2023 🗣️ Talk on “Introduction to Deep Reinforcement Learning” at the University of Pretoria and Indaba X Ghana.
Apr, 2023 🌟 Attended ICLR in Kigali, Rwanda.
Aug, 2022 🛠️ Co-Organiser of the ML Efficiency Workshop at the DLI.
Aug, 2022 🛠️ Programme committee member and Practicals Chair of Deep Learning Indaba (DLI)notebooks 2022, ML Prac, RL Prac.
Jun, 2022 🗣️ Taught an “Introduction to Machine Learning” course at Africa to Silicon Valley.
Mar, 2021 🤖 Joined the Multi-Agent RL research team at InstaDeep.
Dec, 2019 🌟 Attended NeurIPS in Vancouver, Canada.
Aug, 2019 🏆 Won Best Poster (1 out of 194) at the Deep Learning Indaba, sponsored by Microsoft.

selected publications

  1. AAMAS
    Probing Dec-POMDP Reasoning in Cooperative MARL
    Kale-ab Abebe Tessera, Leonard Hinckeldey, Riccardo Zamboni, and 2 more authors
    In The 25th International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS), Oral, 2026
  2. AAMAS
    Fairness over Equality: Correcting Social Incentives in Asymmetric Sequential Social Dilemmas
    Alper Demir, Hüseyin Aydın, Kale-ab Abebe Tessera, and 2 more authors
    In The 25th International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS), Oral, 2026
  3. NeurIPS
    HyperMARL: Adaptive Hypernetworks for Multi-Agent RL
    Kale-ab Abebe Tessera, Arrasy Rahman, Amos Storkey, and 1 more author
    In The Thirty-ninth Annual Conference on Neural Information Processing Systems (NeurIPS), 2025
  4. NeurIPS Workshop
    Are we going MAD? Benchmarking Multi-Agent Debate between Language Models for Medical Q&A
    Andries Smit, Paul Duckworth, Nathan Grinsztajn, and 3 more authors
    In Deep Generative Models for Health Workshop NeurIPS 2023, Nov 2023
  5. arXiv
    Mava: a research framework for distributed multi-agent reinforcement learning
    Arnu Pretorius *, Kale-ab Abebe Tessera *, Andries P Smit *, and 8 more authors
    arXiv preprint arXiv:2107.01460v1, Jul 2021