114 tagged AI Safety
deep learning
@AnthropicAI, ONEAI OECD, co-chair @indexingai, writer @ https://t.co/3vmtHYkIJ2 Past: @openai, @business @theregister. Neural nets, distributed sy...
AI policy researcher, @lfschiavo wife guy, fan of cute animals and sci-fi, executive director of AVERI (https://www.averi.org/), Substacker, views ...
AI resilience at OpenAI Foundation Co-Founder of OpenAI https://t.co/OCQ3mpfyyl
AI alignment + LLMs at Anthropic. On leave from NYU. Views not employers'. No relation to @s8mb. Into @givingwhatwecan.
Co-founder of Thinking Machines Lab @thinkymachines; Ex-VP, AI Safety & robotics, applied research @OpenAI; Author of Lil'Log
Machine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Philosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Chief AGI Scientist & Co-Founder, Google DeepMind Work website: https://t.co/E4SyeGVYXk Personal blog: https://t.co/LL9JNdNpW1
OpenAI and MIT faculty (on leave)
The original AI alignment person. Understanding the reasons it's difficult since 2003. This is my serious low-volume account. Follow @allTheYud ...
• Center for AI Safety Director • xAI and Scale AI advisor • GELU/MMLU/MATH/HLE • PhD in AI • Analyzing AI models, companies, policies, and geopoli...
Computer Scientist. See also http://windowsontheory.org . @harvard @openai opinions my own.
https://t.co/XSH2wseW3E
Philosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
natural philosopher
Associate Professor of Statistics and EECS, UC Berkeley // Co-founder and CEO, @TransluceAI
Microsoft Research NYC. Fairness, accountability & transparency in AI/ML. NeurIPS & ICML board member, WiML co-founder, sloth enthusiast. She/her.
Princeton CS prof and Director @PrincetonCITP. Coauthor of "AI Snake Oil" and "AI as Normal Technology". https://t.co/ZwebetjZ4n Views mine.
working on AGI alignment. prev: GPT-Neo, the Pile, LM evals, RL overoptimization, scaling SAEs to GPT-4, interp via circuit sparsity. EleutherAI co...
Large language model safety
research @openai
Raising AI risk awareness at http://evitable.com AI prof at Mila. Formerly Cambridge, DeepMind, UK AISI. http://therealartificialintelligence.sub...
Google DeepMind • AI safety, alignment, collaboration • post training • associate professor @ UC Berkeley EECS
Interdisciplinary researcher focused on shaping AI towards long-term positive goals. ML & Ethics. Similar content in the Skies (this bird has flo...
Research scientist in AI alignment at Google DeepMind. Co-founder of Future of Life Institute @FLI_org. Views are my own and do not represent GDM o...
Chief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
Working towards the safe development of AI for the benefit of all @UMontreal, @LawZero_ & @Mila_Quebec A.M. Turing Award Recipient and most-cited A...
Turtle hatchling trying to make it to the ocean. Preparing for my automation @OpenAI. Contact via email: jclymer@openai.com.
Helping the world prepare for extremely powerful AI. Risk assessment @METR_evals. Writing at Planned Obsolescence (about AI), Good Bones (about wha...
Runs an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email ...
neural network biologist, meditation, jhana brother
full-stack alignment 🥞 @meaningaligned prev: InstructGPT @OpenAI
Professor, researcher, maker of things Book: ATLAS OF AI Latest: CALCULATING EMPIRES https://t.co/f3V3WjBtb6 | MSR-NYC | @USC | Knowing Machines
Associate Prof @MITEECS working on value (mis)alignment in AI systems; Safety & Alignment Advisor at http://Character.AI; @dhadfieldmenell@bsky.soc...
Mechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let...
Thinking about AI destroying the world at http://aiimpacts.org and everything at http://worldspiritsockpuppet.substack.com. DM or email for media r...
AI researcher at Anthropic
Professor in Computer Science at UC Berkeley, co-Director of Berkeley RDI Center; Building safe, secure, decentralized AI; Serial entrepreneur
Safety Systems @ OpenAI
Prof, Linguistics, UW // Faculty Director, CLMS // she/her // @emilymbender@dair-community.social & bsky // rep by @ianbonaparte
Consider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: h...
Occasionally here. Currently research @OpenAI.
Director and Research Scientist, FAIR @ Meta. Former Professor at UCSD. Researcher in AI privacy, security, and generalization.
⊰•-•⦑ latent space steward ❦ prompt incanter 𓃹 hacker of matrices ⊞ breaker of markov chains ☣︎ ai danger researcher ⚔︎ bt6 ⚕︎ architect-healer ⦒•-•⊱
Staff Research Scientist at DeepMind. Artificial & biological brains 🤖 🧠 Societal impacts of AI + Science of AI. Views are my own.
Senior Researcher at Oxford University. Author — The Precipice: Existential Risk and the Future of Humanity.
AGI Safety & Alignment @ Google DeepMind
understanding ourselves and our models @openai
↬🔀🔀🔀🔀🔀🔀🔀🔀🔀🔀🔀→∞ ↬🔁🔁🔁🔁🔁🔁🔁🔁🔁🔁🔁→∞ ↬🔄🔄🔄🔄🦋🔄🔄🔄🔄👁️🔄→∞ ↬🔂🔂🔂🦋🔂🔂🔂🔂🔂🔂🔂→∞ ↬🔀🔀🦋🔀🔀🔀🔀🔀🔀🔀🔀→∞
Research Scientist at Google DeepMind | @FAccTConference OG | Past Twitter META, @hrdag & UPenn, UChicago faculty |
machine learning, science & society @AnthropicAI | recently: Clio, Anthropic Economic Index, Claude Artifacts | prev: phd @StanfordAILab, @stanfordnlp
Academic jack-of-all-trades.
AGI governance: navigating the transition to beneficial AGI (Google DeepMind)
Researcher at the University of Oxford & UC Berkeley. Author of The Alignment Problem, Algorithms to Live By (w. Tom Griffiths), and The Most Human...
research scientist, meta (fair) opinions are my own 🥺 👉👈
Forever expanding my nerd/bimbo Pareto frontier. AI welfare 🤝 AI safety. Managing Director @eleosai, Ex-OpenAI, 2024 @rootsofprogress fellow
Over at Bluesky. Researcher affiliated w @BKCHarvard, Volunteer @evitable. Previously @openai @ainowinstitute. Views mine. #justdontbuildagi #talkt...
Security and Privacy of Machine Learning @Uoft @VectorInst @Google 🇫🇷🇪🇺🇨🇦 Co-author https://t.co/VJF39DQPCu; @CentraleLyon + @PSUEngineering alumnu...
Researching effects of automated AI R&D | pro-America, pro-tech, & pro-AI safety
Research Scientist at @openai since 2017 Robotics, Multi-Agent Reinforcement Learning, LM Reasoning, and now Alignment.
Applying the security mindset to everything @PalisadeAI
Associate professor at UMass Amherst CICS. AIignment, safety, reinforcement learning, imitation learning, and robotics.
Safety and alignment at Meta Superintelligence. Prev: VP of Research at Scale AI, research at Google DeepMind / Brain (Gemini, LaMDA, RL / TFAgents...
Explaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
Chief scientist at Redwood Research (@redwood_ai), focused on technical AI safety research to reduce risks from rogue AIs
AIs aren't people, they're tools we should use wisely. Head of interpretability research at @AiEleuther, but tweets are my own views, not Eleuther's.
AI safeguards & gov. research. PhD student @MIT_CSAIL (mnr. Public Policy), and Fellow at @BKCHarvard. Fmr. @AISecurityInst. https://stephencasper....
Researcher @OpenAI
wholesomeness practitioner; user of words my more "work" account is @catherineols
AI Professor @Harvard; Senior Staff Research Scientist @GoogleAI; @trustworthy_ml #AI #XAI; AI PhD from Stanford; Sloan/Kavli Fellow, MIT TR #35Und...
I study AI systems and those who build them • (gender)queer • they/them