Amanda Askell
@AmandaAskell
AI SAFETYPhilosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Joshua Achiam
@jachiam0
AI SAFETYFreedom, flourishing, and abundance. Chief Futurist @openai. Main author of http://spinningup.openai.com
Catherine Olsson
@catherineols
RESEARCH ENGINEERHanging out with Claude, improving its behavior, and building tools to support that @AnthropicAI 😁 prev: @open_phil @googlebrain @openai (@microcovid)
Riley Goodside
@goodside
ENGINEERScreenshots of chatbots since 2022. Formerly: Google DeepMind, Scale.
Graham Neubig
@gneubig
ACADEMICAssociate professor @LTIatCMU. Co-founder/chief scientist @OpenHandsDev. I mostly work on modeling language.
Leo Gao
@nabla_theta
AI SAFETYworking on AGI alignment. prev: GPT-Neo, the Pile, LM evals, RL overoptimization, scaling SAEs to GPT-4, interp via circuit sparsity. EleutherAI cofounder.
Joe Carlsmith
@jkcarlsmith
AI SAFETYPhilosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
kipply
@kipperrii
RESEARCH ENGINEER"uncanny ability to be mentioned in every slack thread about code that's mysteriously breaking" - claude
Owain Evans
@OwainEvans_UK
AI SAFETYRuns an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email to DM.
Ajeya Cotra
@ajeya_cotra
AI SAFETYHelping the world prepare for extremely powerful AI. Risk assessment @METR_evals. Writing at Planned Obsolescence (about AI), Good Bones (about whatever).
Neel Nanda
@NeelNanda5
AI SAFETYMechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let's go do it!
Buck Shlegeris
@bshlgrs
AI SAFETYCEO@Redwood Research (@redwood_ai), working on technical research to reduce catastrophic risk from AI misalignment. bshlegeris@gmail.com
⿻ Andrew Trask
@iamtrask
AI SAFETYi teach AI on X building AI with attribution-based control @openminedorg, @GoogleDeepMind, @OxfordUni, @UN, @GovAIOrg, and @CFR_org
Katja Grace 🔍
@KatjaGrace
AI SAFETYThinking about AI destroying the world at http://aiimpacts.org and everything at http://worldspiritsockpuppet.substack.com. DM or email for media requests.
Alexander Berger
@albrgr
FOUNDEREnjoys a good applied micro paper. CEO of @coeff_giving. Views my own, tweets self-destruct every once in a while.
William MacAskill
@willmacaskill
AI SAFETYConsider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: http://forethought.org
Rob Wiblin
@robertwiblin
CREATORHost of the 80,000 Hours Podcast. Exploring the inviolate sphere of ideas one interview at a time: http://80000hours.org/podcast/
Connor Leahy
@NPCollapse
POLICYUS Director @ControlAI - Leave me anonymous feedback: http://bit.ly/3RZbu7x - I don't know how to save the world, but dammit I'm gonna try
Yo Shavit
@yonashav
AI SAFETYpolicy for v smart things @openai. Past: CS PhD @HarvardSEAS/@SchmidtFutures/@MIT_CSAIL. Tweets my own; on my head be it.
Allan Dafoe
@AllanDafoe
AI SAFETYAGI governance: navigating the transition to beneficial AGI (Google DeepMind)
Rosie Campbell
@RosieCampbell
AI SAFETYForever expanding my nerd/bimbo Pareto frontier. AI welfare 🤝 AI safety. Managing Director @eleosai, Ex-OpenAI, 2024 @rootsofprogress fellow
Adam Gleave
@ARGleave
AI SAFETYCEO & co-founder @FARAIResearch non-profit | PhD from @berkeley_ai | Alignment & robustness | on bsky as http://gleave.me
Kayo Yin
@kayo_yin
RESEARCHERPhD student @berkeley_ai. AI persuasion, safety, sign language. Prev @carnegiemellon @polytechnique, intern @msftresearch @deepmind. 🇫🇷🇯🇵
Daniel Eth (yes, Eth is my actual last name)
@daniel_271828
AI SAFETYResearching effects of automated AI R&D | pro-America, pro-tech, & pro-AI safety
Rob Miles
@robertskmiles
AI SAFETYExplaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
uncatherio
@uncatherio
AI SAFETYwholesomeness practitioner; user of words my more "work" account is @catherineols
Daniel Ek
@eldsjal
FOUNDERFather, Principal @ PM - Trying to build things that matter | Founder of Spotify and Neko Health
Steven Adler
@sjgadler
AI SAFETYAI safety researcher (ex-OpenAI: danger evals, AGI readiness, etc), writing at https://clear-eyed.ai
Eli Lifland
@eli_lifland
AI SAFETYAI forecasting and governance @AI_Futures_. Co-author of AI 2027 and the AI Futures Model. Also @aidigest_, @SamotsvetyF. Prev @oughtinc
Marius Hobbhahn
@MariusHobbhahn
AI SAFETYCEO at Apollo Research @apolloaievals prev. ML PhD with Philipp Hennig & AI forecasting @EpochAIResearch
Peter Wildeford🇺🇸🚀
@peterwildeford
POLICYHead of Policy @AIPolicyNetwork. Ohioan. World champion forecaster. Former data scientist. Protecting liberty and prosperity in the age of superintelligence.
Joel Lehman
@joelbot3000
RESEARCHERML researcher, co-author Why Greatness Cannot Be Planned. Creative+safe AI, philosophy. SiR @second_natureai; prev OpenAI / Uber AI / Geometric Intelligence
Rob Bensinger ⏹️
@robbensinger
CREATORComms @MIRIBerkeley. RT = increased vague psychological association between myself and the tweet.