Upcoming Event

On Finding What You’re (Not) Looking For: Prospects and Challenges For AI-Driven Discovery

Speaker: Dr André Curtis-Trudel, Lingnan University Abstract:  Recent high-profile scientific achievements by machine learning (ML) and especially deep learning (DL) systems have reinvigorated interest in ML for automated scientific discovery (e.g., Wang et al. 2023). The hope, in rough outline, is that ML or DL might be used to identify promising novel events, phenomena, hypotheses, …

On Finding What You’re (Not) Looking For: Prospects and Challenges For AI-Driven Discovery Read More »

Conditional Analysis of Model Abilities

Speaker: Jacqueline Harding, Stanford University Abstract:  What can contemporary machine learning (ML) models do? Given the proliferation of ML models in society, answering this question matters to a variety of stakeholders, both public and private. The evaluation of models’ capabilities is rapidly emerging as a key subfield of modern ML, buoyed by regulatory attention and …

Conditional Analysis of Model Abilities Read More »

Does the No Miracles Argument Apply to AI?

Speaker: Prof Darrell Rowbottom, Lingnan University Abstract:  According to the standard no miracles argument, science’s predictive success is best explained by the approximate truth of its theories. In contemporary science, however, machine learning systems, such as AlphaFold2, are also remarkably predictively successful. Thus, we might ask what best explains such successes. Might these AIs accurately …

Does the No Miracles Argument Apply to AI? Read More »

Measuring Capabilities and Safety

Speaker: Dr Dan Hendrycks, Centre for AI Safety Abstract Dr Hendrycks will discuss principles for measuring capabilities of AI systems, and walk through popular general capabilities benchmarks. He will then discuss how general capabilities can be separated from the measurement of their safety, then overview new ways to measure safety.

Regulation by Benchmark

Speaker: Dr Peter Salib, The University of Houston Abstract:  Assume that we succeed in crafting effective safety benchmarks for frontier AI systems. By “effective,” I mean benchmarks that are both aimed at measuring the riskiest capabilities and able to reliably measure them. It would then seem sensible to integrate those benchmarks into safety laws governing …

Regulation by Benchmark Read More »

Evaluating AI Systems for Moral Patienthood

Speaker: Rob Long, Center for AI Safety Abstract:  AI systems could plausibly deserve moral consideration in the not-too-distant future. Although precise evaluations are difficult to devise in light of moral and empirical uncertainty about AI moral patienthood, they will be an important tool for handling this issue responsibly and for avoiding both under- and over-attributing …

Evaluating AI Systems for Moral Patienthood Read More »

Scroll to Top