David Duvenaud
@DavidDuvenaud
AI SAFETYMachine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Matthew Johnson
@SingularMattrix
RESEARCH ENGINEERResearcher at Google Brain. I work on JAX (https://github.com/google/jax).
Catherine Olsson
@catherineols
RESEARCH ENGINEERHanging out with Claude, improving its behavior, and building tools to support that @AnthropicAI 😁 prev: @open_phil @googlebrain @openai (@microcovid)
Boaz Barak
@boazbaraktcs
AI SAFETYComputer Scientist. See also http://windowsontheory.org . @harvard @openai opinions my own.
Riley Goodside
@goodside
ENGINEERScreenshots of chatbots since 2022. Formerly: Google DeepMind, Scale.
Leo Gao
@nabla_theta
AI SAFETYworking on AGI alignment. prev: GPT-Neo, the Pile, LM evals, RL overoptimization, scaling SAEs to GPT-4, interp via circuit sparsity. EleutherAI cofounder.
Joe Carlsmith
@jkcarlsmith
AI SAFETYPhilosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
Victoria Krakovna
@vkrakovna
AI SAFETYResearch scientist in AI alignment at Google DeepMind. Co-founder of Future of Life Institute @FLI_org. Views are my own and do not represent GDM or FLI.
Owain Evans
@OwainEvans_UK
AI SAFETYRuns an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email to DM.
Dylan HadfieldMenell
@dhadfieldmenell
AI SAFETYAssociate Prof @MITEECS working on value (mis)alignment in AI systems; Safety & Alignment Advisor at http://Character.AI; @dhadfieldmenell@bsky.social; he/him
Theo Weber
@theophaneweber
RESEARCH ENGINEERResearch scientist @ DeepMind; currently working on thinking/reasoning in Gemini.
Buck Shlegeris
@bshlgrs
AI SAFETYCEO@Redwood Research (@redwood_ai), working on technical research to reduce catastrophic risk from AI misalignment. bshlegeris@gmail.com
Stephanie Chan
@scychan_brains
AI SAFETYStaff Research Scientist at DeepMind. Artificial & biological brains 🤖 🧠 Societal impacts of AI + Science of AI. Views are my own.
Stella Biderman @ ICLR
@BlancheMinerva
OPEN SOURCEEnsuring that tech companies don't have a monopoly on being able to do research on cutting edge AI @AiEleuther. She/her
William MacAskill
@willmacaskill
AI SAFETYConsider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: http://forethought.org
Connor Leahy
@NPCollapse
POLICYUS Director @ControlAI - Leave me anonymous feedback: http://bit.ly/3RZbu7x - I don't know how to save the world, but dammit I'm gonna try
Alex Tamkin
@AlexTamkin
AI SAFETYmachine learning, science & society @AnthropicAI | recently: Clio, Anthropic Economic Index, Claude Artifacts | prev: phd @StanfordAILab, @stanfordnlp
Yo Shavit
@yonashav
AI SAFETYpolicy for v smart things @openai. Past: CS PhD @HarvardSEAS/@SchmidtFutures/@MIT_CSAIL. Tweets my own; on my head be it.
Alexey Guzey
@alexeyguzey
CREATORhttp://guzey.com, http://newscience.org, @openai; in pursuit of a just, beautiful future.
Brendan O'Donoghue
@bodonoghue85
RESEARCH ENGINEERDirector of Research at @GoogleDeepMind, working on generative models, deep learning, RL. PhD from @stanford. Gemini Diffusion lead.
Adam Gleave
@ARGleave
AI SAFETYCEO & co-founder @FARAIResearch non-profit | PhD from @berkeley_ai | Alignment & robustness | on bsky as http://gleave.me
Brian Christian
@brianchristian
AI SAFETYResearcher at the University of Oxford & UC Berkeley. Author of The Alignment Problem, Algorithms to Live By (w. Tom Griffiths), and The Most Human Human.
Ahmad Beirami
@abeirami
RESEARCH ENGINEERstealth // ex Gemini RL+Inference @GoogleDeepMind // Chat AI @AIatMeta // RL Agents @EA // ML+Information Theory @MIT+@Harvard+@GeorgiaTech
David Abel
@dabelcs
RESEARCHERScientist @GoogleDeepMind and Fellow @EdinburghUni | RL, philosophy, agency, foundations, alignment | office hours: http://tinyurl.com/dabel-mtg
Max Kleiman-Weiner
@maxhkw
ACADEMICprofessor @UW scientist @GoogleDeepMind. computational cognitive models of social minds and machines. priors: PhD @MIT founder @diffeo @CSM_ai
Brandon McKinzie
@mckbrando
RESEARCH ENGINEERresearch @OpenAI | prev: multimodal @Apple, physics/cs @UCBerkeley
Niloofar
@niloofar_mire
RESEARCHERTechnical staff @humansand, incoming asst. prof @LTIatCMU @CMU_EPP, ex RS in @AIatMeta, postdoc @uwcse, Ph.D. @ucsd_cse, former @MSFTResearch -Privacy, ML, NLP
Rylan Schaeffer
@RylanSchaeffer
RESEARCHERAI RS @ Meta TBD. On-Leave from Stanford w/ @sanmikoyejo. Prev @ Gemini, Meta, MIT, Harvard, Uber, UCL, UC Davis
xuan (ɕɥɛn / sh-yen)
@xuanalogue
AI SAFETYAssistant Professor at NUS. Scaling cooperation for an increasingly automated future. PhD @ MIT ProbComp / CoCoSci. Pronouns: 祂/伊
Nora Belrose
@norabelrose
AI SAFETYAIs aren't people, they're tools we should use wisely. Head of interpretability research at @AiEleuther, but tweets are my own views, not Eleuther's.
Rob Miles
@robertskmiles
AI SAFETYExplaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
Eli Lifland
@eli_lifland
AI SAFETYAI forecasting and governance @AI_Futures_. Co-author of AI 2027 and the AI Futures Model. Also @aidigest_, @SamotsvetyF. Prev @oughtinc