Sam Bowman
@sleepinyourhat
AI SAFETYAI alignment + LLMs at Anthropic. On leave from NYU. Views not employers'. No relation to @s8mb. Into @givingwhatwecan.
David Duvenaud
@DavidDuvenaud
ACADEMICMachine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Amanda Askell
@AmandaAskell
AI SAFETYPhilosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Catherine Olsson
@catherineols
RESEARCH ENGINEERHanging out with Claude, improving its behavior, and building tools to support that @AnthropicAI 😁 prev: @open_phil @googlebrain @openai (@microcovid)
Neel Nanda
@NeelNanda5
AI SAFETYMechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let's go do it!
Owain Evans
@OwainEvans_UK
AI SAFETYRuns an AI Safety research group in Berkeley (Truthful AI) + Affiliate at UC Berkeley. Past: Oxford Uni, TruthfulQA, Reversal Curse. Prefer email to DM.
Geoffrey Irving
@geoffreyirving
AI SAFETYChief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
kipply
@kipperrii
RESEARCH ENGINEER"uncanny ability to be mentioned in every slack thread about code that's mysteriously breaking" - claude
Dylan HadfieldMenell
@dhadfieldmenell
AI SAFETYAssociate Prof @MITEECS working on value (mis)alignment in AI systems; Safety & Alignment Advisor at http://Character.AI; @dhadfieldmenell@bsky.social; he/him
Chelsea Sierra Voss
@csvoss
RESEARCH ENGINEERengineeress ✨ Member of Technical Staff @openai // board $QCLS // past @pilothq, @sendwaveapp, @khanacademy, CS/math @MIT, 2x IBO gold
Kevin Roose
@kevinroose
CREATORNYT tech columnist, Hard Fork co-host, high-perplexity language model. Writing a book about AGI.
Ajeya Cotra
@ajeya_cotra
AI SAFETYHelping the world prepare for extremely powerful AI. Risk assessment @METR_evals. Writing at Planned Obsolescence (about AI), Good Bones (about whatever).
Yo Shavit
@yonashav
AI SAFETYpolicy for v smart things @openai. Past: CS PhD @HarvardSEAS/@SchmidtFutures/@MIT_CSAIL. Tweets my own; on my head be it.
William MacAskill
@willmacaskill
RESEARCHERConsider donating 10% to effective charities: http://www.givingwhatwecan.org/pledge Or a career for impact: http://80000hours.org My research: http://forethought.org
Sayash Kapoor
@sayashk
RESEARCHERI tweet about AI agents, AI evals, AI for science. AI as Normal Technology: http://bit.ly/ai-nt Book: http://bit.ly/ai-so
xuan (ɕɥɛn / sh-yen)
@xuanalogue
ACADEMICAssistant Professor at NUS. Scaling cooperation for an increasingly automated future. PhD @ MIT ProbComp / CoCoSci. Pronouns: 祂/伊
Andy Matuschak
@andy_matuschak
RESEARCHERMore wonder, more insight, more expression, more joy! R&D on tools for thought, with a focus on memory, reading, and attention. Currently working on @pico.
Reiner Pope
@reinerpope
FOUNDERCEO and founder, @MatXComputing, developing high throughput chips tailored for LLMs
Joe Carlsmith
@jkcarlsmith
AI SAFETYPhilosophy, futurism, AI. Working on Claude's values @AnthropicAI. Formerly @coeff_giving. Opinions my own.
Gretchen Krueger
@GretchenMarina
AI SAFETYOver at Bluesky. Researcher affiliated w @BKCHarvard, Volunteer @evitable. Previously @openai @ainowinstitute. Views mine. #justdontbuildagi #talktomenotchatgpt
Nathan Labenz
@labenz
CREATORAI Scout, building text-2-video @Waymark, host of The Cognitive Revolution podcast
Rob Wiblin
@robertwiblin
CREATORHost of the 80,000 Hours Podcast. Exploring the inviolate sphere of ideas one interview at a time: http://80000hours.org/podcast/
Adam Gleave
@ARGleave
AI SAFETYCEO & co-founder @FARAIResearch non-profit | PhD from @berkeley_ai | Alignment & robustness | on bsky as http://gleave.me
Rosie Campbell
@RosieCampbell
AI SAFETYForever expanding my nerd/bimbo Pareto frontier. AI welfare 🤝 AI safety. Managing Director @eleosai, Ex-OpenAI, 2024 @rootsofprogress fellow
Allan Dafoe
@AllanDafoe
AI SAFETYAGI governance: navigating the transition to beneficial AGI (Google DeepMind)
rohit
@krishnanrohit
CREATOREssays: http://www.strangeloopcanon.com | Book: http://amazon.com/dp/B0CJ9F327M | World model: https://github.com/Strange-Lab-AI/vei
Marius Hobbhahn
@MariusHobbhahn
AI SAFETYCEO at Apollo Research @apolloaievals prev. ML PhD with Philipp Hennig & AI forecasting @EpochAIResearch
Ashvin Nair
@ashvinair
RESEARCH ENGINEERRL foundations @cursor_ai. Prev: o1, o3, Code Interpreter @openai, 9 years learning to poke by poking at UC Berkeley
Katja Grace 🔍
@KatjaGrace
AI SAFETYThinking about AI destroying the world at http://aiimpacts.org and everything at http://worldspiritsockpuppet.substack.com. DM or email for media requests.
Ben Kuhn
@benkuhn
RESEARCH ENGINEERCare a lot and try hard • making language models safer @AnthropicAI • prev CTO @WaveSenegal 🐧❤️
Steven Adler
@sjgadler
AI SAFETYAI safety researcher (ex-OpenAI: danger evals, AGI readiness, etc), writing at https://clear-eyed.ai
Allen Nie (🇺🇦☮️)
@allenainie
RESEARCH ENGINEERGemini training @GoogleDeepMind. Working on RL. Co-creator of Trace. Prev: RL PhD @StanfordAILab, @MSFTResearch, @DeepMind, @AWS Neuron
Matthew Barnett
@MatthewJBar
FOUNDERCo-founder of @MechanizeWork Married to @natalia__coelho email: matthew at mechanize dot work
Rob Miles
@robertskmiles
AI SAFETYExplaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
Eli Lifland
@eli_lifland
RESEARCHERAI forecasting and governance @AI_Futures_. Co-author of AI 2027 and the AI Futures Model. Also @aidigest_, @SamotsvetyF. Prev @oughtinc
Seán Ó hÉigeartaigh
@S_OhEigeartaigh
AI SAFETYDirector of http://ai-far.org at Uni of Cambridge | Researching Big Risks, and impacts of AI & emerging tech. Opinions own
Liv
@livgorton
AI SAFETY✨ asking sand to show its work // currently @AnthropicAI, prev founding RS @GoodfireAI // creating a more beautiful future