Computing Superior Counter-Examples for Conformant Planning

In a counterexample based approach to conformant planning, choosing the right counterexample can improve performance. We formalise this observation by introducing the notion of “superiority” of a counterexample over another one,that holds whenever the superior counterexample exhibits more tags than the latter. We provide a theoretical explanation that supports the strategy of searching for maximally superior counterexamples, and we show how this strategy can be implemented. The empirical experiments validate our approach.

Read More
Attributions of Ethical Responsibility by Artificial Intelligence Practitioners

Through in-depth interviews with AI practitioners in Australia, this paper examines perceptions of accountability and responsibility among those who make autonomous systems. We find that AI practitioners envision themselves as mediating technicians, enacting others’ high-level plans and then relinquishing control of the products they produce. Findings highlight “ethics” in AI as a challenge that distributes among complex webs of human and mechanized subjects.

Read More
Military Applications of AI, International Security, and Arms Control Workshop

Professor Toni Erskine, HMI Discovery Lead, presented at the workshop on 'Military Applications of AI, International Security, and Arms Control', hosted by the United Nations Institute for Disarmament Research, convened by David Danks (Carnegie Mellon University), Paul Meyer (Simon Fraser University), and Giaocomo Paoli (UNIDIR). The workshop was held on the 30th and 31st of January 2020 in Santa Monica, California.

Read More
Modelling Information Cascades with Self-Exciting Processes via Generalized Epidemic Models

Epidemic models and self-exciting processes are two types of models used to describe diffusion phenomena online and offline. These models were originally developed in different scientific communities, and their commonalities are under-explored. This work establishes, for the first time, a general connection between the two model classes via three new mathematical components.

Read More
Mathematical decisions and non-causal elements of explainable AI

I offer a multi-faceted conceptual framework for the explanation and the interpretations of algorithmic decisions, and I claim that this framework can lay the groundwork for a focused discussion among multiple stakeholders about the social implications of algorithmic decision-making, as well as AI governance and ethics more generally.

Read More
AI for Social Good

From 2019, Professor Toni Erskine, Discovery Lead, has served on the Advisory Group for the Google/United Nations Economic and Social Commission for Asia and the Pacific (ESCAP), ‘AI for the Social Good’ Research Network, administered by the Association for Pacific Rim Universities.

Read More
News, ResourcesHMI StaffToni