•  18
    Social Choice for AI Alignment: Dealing with Diverse Human Feedback
    with Vincent Conitzer, Rachel Freedman, Jobst Heitzig, Wesley H. Holliday, Bob M. Jacobs, Nathan Lambert, Eric Pacuit, Stuart Russell, Hailey Schoelkopf, Emanuel Tewolde, and William S. Zwicker
    Foundation models such as GPT-4 are fine-tuned to avoid unsafe or otherwise problematic behavior, so that, for example, they refuse to comply with requests for help with committing crimes or with producing racist text. One approach to fine-tuning, called reinforcement learning from human feedback, learns from humans' expressed preferences over multiple outputs. Another approach is constitutional AI, in which the input from humans is a list of high-level principles. But how do we deal with potent…Read more
  •  54
    Probing the quantitative–qualitative divide in probabilistic reasoning
    with Duligur Ibeling, Thomas Icard, and Krzysztof Mierzewski
    Annals of Pure and Applied Logic. forthcoming.
  •  341
    Is Causal Reasoning Harder Than Probabilistic Reasoning?
    with Duligur Ibeling and Thomas Icard
    Review of Symbolic Logic 17 (1): 106-131. 2024.
    Many tasks in statistical and causal inference can be construed as problems of entailment in a suitable formal language. We ask whether those problems are more difficult, from a computational perspective, for causal probabilistic languages than for pure probabilistic (or “associational”) languages. Despite several senses in which causal reasoning is indeed more complex—both expressively and inferentially—we show that causal entailment (or satisfiability) problems can be systematically and robust…Read more