111 tagged AI Safety
deep learning
AI policy researcher, @lfschiavo wife guy, fan of cute animals and sci-fi, executive director of AVERI (https://www.averi.org/), Substacker, views ...
@AnthropicAI, ONEAI OECD, co-chair @indexingai, writer @ http://importai.net Past: @openai, @business @theregister. Neural nets, distributed system...
Co-founder of Thinking Machines Lab @thinkymachines; Ex-VP, AI Safety & robotics, applied research @OpenAI; Author of Lil'Log
AI resilience at OpenAI Foundation Co-Founder of OpenAI https://woj.world
AI alignment + LLMs at Anthropic. On leave from NYU. Views not employers'. No relation to @s8mb. Into @givingwhatwecan.
Machine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Philosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Chief AGI Scientist & Co-Founder, Google DeepMind Work website: http://www.deepmind.com Personal blog: http://www.vetta.org
OpenAI and MIT faculty (on leave)
The original AI alignment person. Understanding the reasons it's difficult since 2003. This is my serious low-volume account. Follow @allTheYud ...
• Center for AI Safety Director • xAI and Scale AI advisor • GELU/MMLU/MATH/HLE • PhD in AI • Analyzing AI models, companies, policies, and geopoli...
Computer Scientist. See also http://windowsontheory.org . @harvard @openai opinions my own.
Alignment team lead at Anthropic
Microsoft Research NYC. Fairness, accountability & transparency in AI/ML. NeurIPS & ICML board member, WiML co-founder, sloth enthusiast. She/her.
natural philosopher
http://situational-awareness.ai
Associate Professor of Statistics and EECS, UC Berkeley // Co-founder and CEO, @TransluceAI
Interdisciplinary researcher focused on shaping AI towards long-term positive goals. ML & Ethics. Similar content in the Skies (this bird has flo...
Google DeepMind • AI safety, alignment, collaboration • post training • associate professor @ UC Berkeley EECS
Princeton CS prof and Director @PrincetonCITP. Coauthor of "AI Snake Oil" and "AI as Normal Technology". https://www.normaltech.ai/ Views mine.
Raising AI risk awareness at http://evitable.com AI prof at Mila. Formerly Cambridge, DeepMind, UK AISI. http://therealartificialintelligence.sub...
Chief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
research @openai
working on AGI alignment. prev: GPT-Neo, the Pile, LM evals, RL overoptimization, scaling SAEs to GPT-4, interp via circuit sparsity. EleutherAI co...
Working towards the safe development of AI for the benefit of all @UMontreal, @LawZero_ & @Mila_Quebec A.M. Turing Award Recipient and most-cited A...
Philosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
Research scientist in AI alignment at Google DeepMind. Co-founder of Future of Life Institute @FLI_org. Views are my own and do not represent GDM o...
Runs an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email ...
Helping the world prepare for extremely powerful AI. Risk assessment @METR_evals. Writing at Planned Obsolescence (about AI), Good Bones (about wha...
Professor, researcher, maker of things Book: ATLAS OF AI Latest: CALCULATING EMPIRES https://calculatingempires.net | MSR-NYC | @USC | Knowing M...
neural network biologist, meditation, jhana brother
Associate Prof @MITEECS working on value (mis)alignment in AI systems; Safety & Alignment Advisor at http://Character.AI; @dhadfieldmenell@bsky.soc...
Mechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let...
full-stack alignment 🥞 @meaningaligned prev: InstructGPT @OpenAI
Prof, Linguistics, UW // Faculty Director, CLMS // she/her // @emilymbender@dair-community.social & bsky // rep by @ianbonaparte
Staff Research Scientist at DeepMind. Artificial & biological brains 🤖 🧠 Societal impacts of AI + Science of AI. Views are my own.
Thinking about AI destroying the world at http://aiimpacts.org and everything at http://worldspiritsockpuppet.substack.com. DM or email for media r...
understanding ourselves and our models @openai
Professor in Computer Science at UC Berkeley, co-Director of Berkeley RDI Center; Building safe, secure, decentralized AI; Serial entrepreneur
AI researcher at Anthropic
Consider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: h...
machine learning, science & society @AnthropicAI | recently: Clio, Anthropic Economic Index, Claude Artifacts | prev: phd @StanfordAILab, @stanfordnlp
⊰•-•⦑ latent space steward ❦ prompt incanter 𓃹 hacker of matrices ⊞ breaker of markov chains ☣︎ ai danger researcher ⚔︎ bt6 ⚕︎ architect-healer ⦒•-•⊱
Turtle hatchling trying to make it to the ocean. Preparing for my automation @OpenAI. Contact via email: jclymer@openai.com.
Research Scientist at Google DeepMind | @FAccTConference OG | Past Twitter META, @hrdag & UPenn, UChicago faculty |
AGI Safety & Alignment @ Google DeepMind
Director and Research Scientist, FAIR @ Meta. Former Professor at UCSD. Researcher in AI privacy, security, and generalization.
Academic jack-of-all-trades.
Senior Researcher at Oxford University. Author — The Precipice: Existential Risk and the Future of Humanity.
↬🔀🔀🔀🔀🔀🔀🔀🔀🔀🔀🔀→∞ ↬🔁🔁🔁🔁🔁🔁🔁🔁🔁🔁🔁→∞ ↬🔄🔄🔄🔄🦋🔄🔄🔄🔄👁️🔄→∞ ↬🔂🔂🔂🦋🔂🔂🔂🔂🔂🔂🔂→∞ ↬🔀🔀🦋🔀🔀🔀🔀🔀🔀🔀🔀→∞
AGI governance: navigating the transition to beneficial AGI (Google DeepMind)
Safety Systems @ OpenAI
Security and Privacy of Machine Learning @Uoft @VectorInst @Google 🇫🇷🇪🇺🇨🇦 Co-author http://cleverhans.io; @CentraleLyon + @PSUEngineering alumnus. ...
Occasionally here. Currently research @OpenAI.
Applying the security mindset to everything @PalisadeAI
Forever expanding my nerd/bimbo Pareto frontier. AI welfare 🤝 AI safety. Managing Director @eleosai, Ex-OpenAI, 2024 @rootsofprogress fellow
research scientist, meta (fair) opinions are my own 🥺 👉👈
Chief scientist at Redwood Research (@redwood_ai), focused on technical AI safety research to reduce risks from rogue AIs
Researcher at the University of Oxford & UC Berkeley. Author of The Alignment Problem, Algorithms to Live By (w. Tom Griffiths), and The Most Human...
Safety and alignment at Meta Superintelligence. Prev: VP of Research at Scale AI, research at Google DeepMind / Brain (Gemini, LaMDA, RL / TFAgents...
OG AGI Cyberneticist @daiosai. Ex AI Safety Team lead @googledeepmind. PhD @CambridgeMLG. Will only post about AI-related stuff, never politics. P...
Trying to make AI go well @AnthropicAI
AI safeguards & gov. research. PhD student @MIT_CSAIL (mnr. Public Policy), and Fellow at @BKCHarvard. Fmr. @AISecurityInst. https://stephencasper....
Member of Technical Staff at Thinking Machines. Human+AI collaboration. Scalable Oversight. Explainability. Prev @AnthropicAI PhD UC Berkeley'25; C...
Over at Bluesky. Researcher affiliated w @BKCHarvard, Volunteer @evitable. Previously @openai @ainowinstitute. Views mine. #justdontbuildagi #talkt...
Head of the Frontier Red Team @anthropicai. 🌎 Make things radically good.
Researching effects of automated AI R&D | pro-America, pro-tech, & pro-AI safety
Assistant Professor at NUS. Scaling cooperation for an increasingly automated future. PhD @ MIT ProbComp / CoCoSci. Pronouns: 祂/伊
AIs aren't people, they're tools we should use wisely. Head of interpretability research at @AiEleuther, but tweets are my own views, not Eleuther's.
Explaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
Associate professor at UMass Amherst CICS. AIignment, safety, reinforcement learning, imitation learning, and robotics.
AI Professor @Harvard; Senior Staff Research Scientist @GoogleAI; @trustworthy_ml #AI #XAI; AI PhD from Stanford; Sloan/Kavli Fellow, MIT TR #35Und...