Upcoming Event

Evaluating AI Agents for Dangerous (Cognitive) Capabilities

Abstract: AI agents based on Large Language Models (LLMs) demonstrate human-level performance at some theory of mind (ToM) tasks (Kosinski 2024; Street et al. 2024). Here ToM is roughly the ability to predict and explain behaviour by attributing mental states to oneself and others. ToM capabilities matter for AI safety because, at least in humans, […]

Evaluating AI Agents for Dangerous (Cognitive) Capabilities Read More »

What remains of the singularity hypothesis?

Abstract: The idea that advances in the cognitive capacities of foundation models like LLMs will lead to a period of rapid, recursive self-improvement — an “intelligence explosion” or “technological singularity” — has recently come under sustained criticism by academic philosophers. I evaluate the extent to which this criticism successfully undermines the argument for a singularity,

What remains of the singularity hypothesis? Read More »

Evaluating LLM Ethical Competence

Abstract: Existing approaches to evaluating LLM ethical competence place too much emphasis on the verdicts—of permissibility and impermissibility—that they render. But ethical competence doesn’t consist in one’s judgments conforming to those of a cohort of crowdworkers. It consists in being able to identify morally relevant features, prioritise among them, associate them with reasons and weave

Evaluating LLM Ethical Competence Read More »

Why ChatGPT Doesn’t Think: An Argument from Rationality (Co-authored with Zhihe Vincent Zhang, ANU)

Abstract: Can AI systems such as ChatGPT think? This paper presents an argument from rationality for the negative answer to this question. The argument is founded on two central ideas. The first is that if ChatGPT thinks, it is not rational, in the sense that it does not respond correctly to its evidence. The second

Why ChatGPT Doesn’t Think: An Argument from Rationality (Co-authored with Zhihe Vincent Zhang, ANU) Read More »

Scroll to Top