Sam Bowman
@sleepinyourhat
AI SAFETYAI alignment + LLMs at Anthropic. On leave from NYU. Views not employers'. No relation to @s8mb. Into @givingwhatwecan.
David Duvenaud
@DavidDuvenaud
AI SAFETYMachine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Amanda Askell
@AmandaAskell
AI SAFETYPhilosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Catherine Olsson
@catherineols
RESEARCH ENGINEERHanging out with Claude, improving its behavior, and building tools to support that @AnthropicAI 😁 prev: @open_phil @googlebrain @openai (@microcovid)
Chelsea Sierra Voss
@csvoss
RESEARCH ENGINEERengineeress ✨ Member of Technical Staff @openai // serious play // past @pilothq, @sendwaveapp, @khanacademy, CS/math @MIT, 2x IBO gold
Geoffrey Irving
@geoffreyirving
AI SAFETYChief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
Joe Carlsmith
@jkcarlsmith
AI SAFETYPhilosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
kipply
@kipperrii
RESEARCH ENGINEER"uncanny ability to be mentioned in every slack thread about code that's mysteriously breaking" - claude
Owain Evans
@OwainEvans_UK
AI SAFETYRuns an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email to DM.
Ajeya Cotra
@ajeya_cotra
AI SAFETYHelping the world prepare for extremely powerful AI. Risk assessment @METR_evals. Writing at Planned Obsolescence (about AI), Good Bones (about whatever).
Dylan HadfieldMenell
@dhadfieldmenell
AI SAFETYAssociate Prof @MITEECS working on value (mis)alignment in AI systems; Safety & Alignment Advisor at http://Character.AI; @dhadfieldmenell@bsky.social; he/him
Neel Nanda
@NeelNanda5
AI SAFETYMechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let's go do it!
Kevin Roose
@kevinroose
CREATORNYT tech columnist, Hard Fork co-host, high-perplexity language model. Writing a book about AGI.
Katja Grace 🔍
@KatjaGrace
AI SAFETYThinking about AI destroying the world at http://aiimpacts.org and everything at http://worldspiritsockpuppet.substack.com. DM or email for media requests.
Alexander Berger
@albrgr
FOUNDEREnjoys a good applied micro paper. CEO of @coeff_giving. Views my own, tweets self-destruct every once in a while.
William MacAskill
@willmacaskill
AI SAFETYConsider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: http://forethought.org
Rob Wiblin
@robertwiblin
CREATORHost of the 80,000 Hours Podcast. Exploring the inviolate sphere of ideas one interview at a time: http://80000hours.org/podcast/
Jason Crawford
@jasoncrawford
CREATORI write and speak about the history & philosophy of progress. Founder, @rootsofprogress. Host of the Progress Conference. Author, The Techno-Humanist Manifesto.
Yo Shavit
@yonashav
AI SAFETYpolicy for v smart things @openai. Past: CS PhD @HarvardSEAS/@SchmidtFutures/@MIT_CSAIL. Tweets my own; on my head be it.
Allan Dafoe
@AllanDafoe
AI SAFETYAGI governance: navigating the transition to beneficial AGI (Google DeepMind)
Rosie Campbell
@RosieCampbell
AI SAFETYForever expanding my nerd/bimbo Pareto frontier. AI welfare 🤝 AI safety. Managing Director @eleosai, Ex-OpenAI, 2024 @rootsofprogress fellow
Adam Gleave
@ARGleave
AI SAFETYCEO & co-founder @FARAIResearch non-profit | PhD from @berkeley_ai | Alignment & robustness | on bsky as http://gleave.me
Andy Matuschak
@andy_matuschak
RESEARCHERMore wonder, more insight, more expression, more joy! R&D on tools for thought, with a focus on memory, reading, and attention. Currently working on @pico.
Gretchen Krueger
@GretchenMarina
AI SAFETYOver at Bluesky. Researcher affiliated w @BKCHarvard, Volunteer @evitable. Previously @openai @ainowinstitute. Views mine. #justdontbuildagi #talktomenotchatgpt
Daniel Eth (yes, Eth is my actual last name)
@daniel_271828
AI SAFETYResearching effects of automated AI R&D | pro-America, pro-tech, & pro-AI safety
xuan (ɕɥɛn / sh-yen)
@xuanalogue
AI SAFETYAssistant Professor at NUS. Scaling cooperation for an increasingly automated future. PhD @ MIT ProbComp / CoCoSci. Pronouns: 祂/伊
Rob Miles
@robertskmiles
AI SAFETYExplaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
uncatherio
@uncatherio
AI SAFETYwholesomeness practitioner; user of words my more "work" account is @catherineols
Steven Adler
@sjgadler
AI SAFETYAI safety researcher (ex-OpenAI: danger evals, AGI readiness, etc), writing at https://clear-eyed.ai
Eli Lifland
@eli_lifland
AI SAFETYAI forecasting and governance @AI_Futures_. Co-author of AI 2027 and the AI Futures Model. Also @aidigest_, @SamotsvetyF. Prev @oughtinc
Tamara Winter
@tamarawinter
CREATORPublisher @stripepress | Board @ifp + @joinFAI | Creator of TACIT, a mini-documentary series following master craftspeople at work
Marius Hobbhahn
@MariusHobbhahn
AI SAFETYCEO at Apollo Research @apolloaievals prev. ML PhD with Philipp Hennig & AI forecasting @EpochAIResearch
Rob Bensinger ⏹️
@robbensinger
CREATORComms @MIRIBerkeley. RT = increased vague psychological association between myself and the tweet.
Ben Kuhn
@benkuhn
RESEARCH ENGINEERCare a lot and try hard • making language models safer @AnthropicAI • prev CTO @WaveSenegal 🐧❤️
Reiner Pope
@reinerpope
FOUNDERCEO and founder, @MatXComputing, developing high throughput chips tailored for LLMs
Stefan Schubert
@StefanFSchubert
RESEARCHERI run The Update newsletter. Book: http://academic.oup.com/book/56384