Effective Altruism News
Effective Altruism News

SUBSCRIBE

RSS Feed
X Feed

ORGANIZATIONS

  • Effective Altruism Forum
  • 80,000 Hours
  • Ada Lovelace Institute
  • Against Malaria Foundation
  • AI Alignment Forum
  • AI Futures Project
  • AI Impacts
  • AI Now Institute
  • AI Objectives Institute
  • AI Safety Camp
  • AI Safety Communications Centre
  • AidGrade
  • Albert Schweitzer Foundation
  • Aligned AI
  • ALLFED
  • Asterisk
  • altLabs
  • Ambitious Impact
  • Anima International
  • Animal Advocacy Africa
  • Animal Advocacy Careers
  • Animal Charity Evaluators
  • Animal Ethics
  • Apollo Academic Surveys
  • Aquatic Life Institute
  • Association for Long Term Existence and Resilience
  • Ayuda Efectiva
  • Berkeley Existential Risk Initiative
  • Bill & Melinda Gates Foundation
  • Bipartisan Commission on Biodefense
  • California YIMBY
  • Cambridge Existential Risks Initiative
  • Carnegie Corporation of New York
  • Center for Applied Rationality
  • Center for Election Science
  • Center for Emerging Risk Research
  • Center for Health Security
  • Center for Human-Compatible AI
  • Center for Long-Term Cybersecurity
  • Center for Open Science
  • Center for Reducing Suffering
  • Center for Security and Emerging Technology
  • Center for Space Governance
  • Center on Long-Term Risk
  • Centre for Effective Altruism
  • Centre for Enabling EA Learning and Research
  • Centre for the Governance of AI
  • Centre for the Study of Existential Risk
  • Centre of Excellence for Development Impact and Learning
  • Charity Entrepreneurship
  • Charity Science
  • Clearer Thinking
  • Compassion in World Farming
  • Convergence Analysis
  • Crustacean Compassion
  • Deep Mind
  • Democracy Defense Fund
  • Democracy Fund
  • Development Media International
  • EA Funds
  • Effective Altruism Cambridge
  • Effective altruism for Christians
  • Effective altruism for Jews
  • Effective Altruism Foundation
  • Effective Altruism UNSW
  • Effective Giving
  • Effective Institutions Project
  • Effective Self-Help
  • Effective Thesis
  • Effektiv-Spenden.org
  • Eleos AI
  • Eon V Labs
  • Epoch Blog
  • Equalize Health
  • Evidence Action
  • Family Empowerment Media
  • Faunalytics
  • Farmed Animal Funders
  • FAST | Animal Advocacy Forum
  • Felicifia
  • Fish Welfare Initiative
  • Fistula Foundation
  • Food Fortification Initiative
  • Foresight Institute
  • Forethought
  • Foundational Research Institute
  • Founders’ Pledge
  • Fortify Health
  • Fund for Alignment Research
  • Future Generations Commissioner for Wales
  • Future of Life Institute
  • Future of Humanity Institute
  • Future Perfect
  • GBS Switzerland
  • Georgetown University Initiative on Innovation, Development and Evaluation
  • GiveDirectly
  • GiveWell
  • Giving Green
  • Giving What We Can
  • Global Alliance for Improved Nutrition
  • Global Catastrophic Risk Institute
  • Global Challenges Foundation
  • Global Innovation Fund
  • Global Priorities Institute
  • Global Priorities Project
  • Global Zero
  • Good Food Institute
  • Good Judgment Inc
  • Good Technology Project
  • Good Ventures
  • Happier Lives Institute
  • Harvard College Effective Altruism
  • Healthier Hens
  • Helen Keller INTL
  • High Impact Athletes
  • HistPhil
  • Humane Slaughter Association
  • IDInsight
  • Impactful Government Careers
  • Innovations for Poverty Action
  • Institute for AI Policy and Strategy
  • Institute for Progress
  • International Initiative for Impact Evaluation
  • Invincible Wellbeing
  • Iodine Global Network
  • J-PAL
  • Jewish Effective Giving Initiative
  • Lead Exposure Elimination Project
  • Legal Priorities Project
  • LessWrong
  • Let’s Fund
  • Leverhulme Centre for the Future of Intelligence
  • Living Goods
  • Long Now Foundation
  • Machine Intelligence Research Institute
  • Malaria Consortium
  • Manifold Markets
  • Median Group
  • Mercy for Animals
  • Metaculus
  • Metaculus | News
  • METR
  • Mila
  • New Harvest
  • Nonlinear
  • Nuclear Threat Initiative
  • One Acre Fund
  • One for the World
  • OpenAI
  • Open Mined
  • Open Philanthropy
  • Organisation for the Prevention of Intense Suffering
  • Ought
  • Our World in Data
  • Oxford Prioritisation Project
  • Parallel Forecast
  • Ploughshares Fund
  • Precision Development
  • Probably Good
  • Pugwash Conferences on Science and World Affairs
  • Qualia Research Institute
  • Raising for Effective Giving
  • Redwood Research
  • Rethink Charity
  • Rethink Priorities
  • Riesgos Catastróficos Globales
  • Sanku – Project Healthy Children
  • Schmidt Futures
  • Sentience Institute
  • Sentience Politics
  • Seva Foundation
  • Sightsavers
  • Simon Institute for Longterm Governance
  • SoGive
  • Space Futures Initiative
  • Stanford Existential Risk Initiative
  • Swift Centre
  • The END Fund
  • The Future Society
  • The Life You Can Save
  • The Roots of Progress
  • Target Malaria
  • Training for Good
  • UK Office for AI
  • Unlimit Health
  • Utility Farm
  • Vegan Outreach
  • Venten | AI Safety for Latam
  • Village Enterprise
  • Waitlist Zero
  • War Prevention Initiative
  • Wave
  • Wellcome Trust
  • Wild Animal Initiative
  • Wild-Animal Suffering Research
  • Works in Progress

PEOPLE

  • Scott Aaronson | Shtetl-Optimized
  • Tom Adamczewski | Fragile Credences
  • Matthew Adelstein | Bentham's Newsletter
  • Matthew Adelstein | Controlled Opposition
  • Vaidehi Agarwalla | Vaidehi's Blog
  • James Aitchison | Philosophy and Ideas
  • Scott Alexander | Astral Codex Ten
  • Scott Alexander | Slate Star Codex
  • Scott Alexander | Slate Star Scratchpad
  • Alexanian & Franz | GCBR Organization Updates
  • Applied Divinity Studies
  • Leopold Aschenbrenner | For Our Posterity
  • Amanda Askell
  • Amanda Askell's Blog
  • Amanda Askell’s Substack
  • Atoms vs Bits
  • Connor Axiotes | Rules of the Game
  • Sofia Balderson | Hive
  • Mark Bao
  • Boaz Barak | Windows On Theory
  • Nathan Barnard | The Good blog
  • Matthew Barnett
  • Ollie Base | Base Rates
  • Simon Bazelon | Out of the Ordinary
  • Tobias Baumann | Cause Prioritization Research
  • Tobias Baumann | Reducing Risks of Future Suffering
  • Nora Belrose
  • Rob Bensinger | Nothing Is Mere
  • Alexander Berger | Marginal Change
  • Aaron Bergman | Aaron's Blog
  • Satvik Beri | Ars Arcana
  • Aveek Bhattacharya | Social Problems Are Like Maths
  • Michael Bitton | A Nice Place to Live
  • Liv Boeree
  • Dillon Bowen
  • Topher Brennan
  • Ozy Brennan | Thing of Things
  • Catherine Brewer | Catherine’s Blog
  • Stijn Bruers | The Rational Ethicist
  • Vitalik Buterin
  • Lynette Bye | EA Coaching
  • Ryan Carey
  • Joe Carlsmith
  • Lucius Caviola | Outpaced
  • Richard Yetter Chappell | Good Thoughts
  • Richard Yetter Chappell | Philosophy, Et Cetera
  • Paul Christiano | AI Alignment
  • Paul Christiano | Ordinary Ideas
  • Paul Christiano | Rational Altruist
  • Paul Christiano | Sideways View
  • Paul Christiano & Katja Grace | The Impact Purchase
  • Evelyn Ciara | Sunyshore
  • Cirrostratus Whispers
  • Jesse Clifton | Jesse’s Substack
  • Peter McCluskey | Bayesian Investor
  • Greg Colbourn | Greg's Substack
  • Ajeya Cotra & Kelsey Piper | Planned Obsolescence
  • Owen Cotton-Barrat | Strange Cities
  • Andrew Critch
  • Paul Crowley | Minds Aren’t Magic
  • Dale | Effective Differentials
  • Max Dalton | Custodienda
  • Saloni Dattani | Scientific Discovery
  • Amber Dawn | Contemplatonist
  • De novo
  • Michael Dello-Iacovo
  • Jai Dhyani | ANOIEAEIB
  • Michael Dickens | Philosophical Multicore
  • Dieleman & Zeijlmans | Effective Environmentalism
  • Anthony DiGiovanni | Ataraxia
  • Kate Donovan | Gruntled & Hinged
  • Dawn Drescher | Impartial Priorities
  • Eric Drexler | AI Prospects: Toward Global Goal Convergence
  • Holly Elmore
  • Sam Enright | The Fitzwilliam
  • Daniel Eth | Thinking of Utils
  • Daniel Filan
  • Lukas Finnveden
  • Diana Fleischman | Dianaverse
  • Diana Fleischman | Dissentient
  • Julia Galef
  • Ben Garfinkel | The Best that Can Happen
  • Matthew Gentzel | The Consequentialist
  • Aaron Gertler | Alpha Gamma
  • Rachel Glennerster
  • Sam Glover | Samstack
  • Andrés Gómez Emilsson | Qualia Computing
  • Nathan Goodman & Yusuf | Longterm Liberalism
  • Ozzie Gooen | The QURI Medley
  • Ozzie Gooen
  • Katja Grace | Meteuphoric
  • Katja Grace | World Spirit Sock Puppet
  • Katja Grace | Worldly Positions
  • Spencer Greenberg | Optimize Everything
  • Milan Griffes | Flight from Perfection
  • Simon Grimm
  • Zach Groff | The Groff Spot
  • Erich Grunewald
  • Marc Gunther
  • Cate Hall | Useful Fictions
  • Chris Hallquist | The Uncredible Hallq
  • Topher Hallquist
  • John Halstead
  • Finn Hambly
  • James Harris | But Can They Suffer?
  • Riley Harris
  • Riley Harris | Million Year View
  • Peter Hartree
  • Shakeel Hashim | Transformer
  • Sarah Hastings-Woodhouse
  • Hayden | Ethical Haydonism
  • Julian Hazell | Julian’s Blog
  • Julian Hazell | Secret Third Thing
  • Jiwoon Hwang
  • Incremental Updates
  • Roxanne Heston | Fire and Pulse
  • Hauke Hillebrandt | Hauke’s Blog
  • Ben Hoffman | Compass Rose
  • Michael Huemer | Fake Nous
  • Tyler John | Secular Mornings
  • Mike Johnson | Open Theory
  • Toby Jolly | Seeking To Be Jolly
  • Holden Karnofsky | Cold Takes
  • Jeff Kaufman
  • Cullen O'Keefe | Jural Networks
  • Daniel Kestenholz
  • Ketchup Duck
  • Oliver Kim | Global Developments
  • Isaac King | Outside the Asylum
  • Petra Kosonen | Tiny Points of Vast Value
  • Victoria Krakovna
  • Ben Kuhn
  • Alex Lawsen | Speculative Decoding
  • Gavin Leech | argmin gravitas
  • Howie Lempel
  • Gregory Lewis
  • Eli Lifland | Foxy Scout
  • Rhys Lindmark
  • Robert Long | Experience Machines
  • Garrison Lovely | Garrison's Substack
  • MacAskill, Chappell & Meissner | Utilitarianism
  • Jack Malde | The Ethical Economist
  • Jonathan Mann | Abstraction
  • Sydney Martin | Birthday Challenge
  • Daniel May
  • Daniel May
  • Conor McCammon | Utopianish
  • Peter McIntyre
  • Peter McIntyre | Conceptually
  • Pablo Melchor
  • Pablo Melchor | Altruismo racional
  • Geoffrey Miller
  • Fin Moorhouse
  • Luke Muehlhauser
  • Neel Nanda
  • David Nash | Global Development & Economic Advancement
  • David Nash | Monthly Overload of Effective Altruism
  • Eric Neyman | Unexpected Values
  • Richard Ngo | Narrative Ark
  • Richard Ngo | Thinking Complete
  • Elizabeth Van Nostrand | Aceso Under Glass
  • Oesterheld, Treutlein & Kokotajlo | The Universe from an Intentional Stance
  • James Ozden | Understanding Social Change
  • Daniel Paleka | AI Safety Takes
  • Ives Parr | Parrhesia
  • Dwarkesh Patel | The Lunar Society
  • Kelsey Piper | The Unit of Caring
  • Michael Plant | Planting Happiness
  • Michal Pokorný | Agenty Dragon
  • Georgia Ray | Eukaryote Writes Blog
  • Ross Rheingans-Yoo | Icosian Reflections
  • Josh Richards | Tofu Ramble
  • Jess Riedel | foreXiv
  • Anna Riedl
  • Hannah Ritchie | Sustainability by Numbers
  • David Roodman
  • Eli Rose
  • Abraham Rowe | Good Structures
  • Siebe Rozendal
  • John Salvatier
  • Anders Sandberg | Andart II
  • William Saunders
  • Joey Savoie | Measured Life
  • Stefan Schubert
  • Stefan Schubert | Philosophical Sketches
  • Stefan Schubert | Stefan’s Substack
  • Nuño Sempere | Measure is unceasing
  • Harish Sethu | Counting Animals
  • Rohin Shah
  • Zeke Sherman | Bashi-Bazuk
  • Buck Shlegeris
  • Jay Shooster | jayforjustice
  • Carl Shulman | Reflective Disequilibrium
  • Jonah Sinick
  • Andrew Snyder-Beattie | Defenses in Depth
  • Ben Snodin
  • Nate Soares | Minding Our Way
  • Kaj Sotala
  • Tom Stafford | Reasonable People
  • Pablo Stafforini | Pablo’s Miscellany
  • Henry Stanley
  • Jacob Steinhardt | Bounded Regret
  • Zach Stein-Perlman | AI Lab Watch
  • Romeo Stevens | Neurotic Gradient Descent
  • Michael Story | Too long to tweet
  • Maxwell Tabarrok | Maximum Progress
  • Max Taylor | AI for Animals
  • Benjamin Todd
  • Brian Tomasik | Reducing Suffering
  • Helen Toner | Rising Tide
  • Philip Trammell
  • Jacob Trefethen
  • Toby Tremlett | Raising Dust
  • Alex Turner | The Pond
  • Alyssa Vance | Rationalist Conspiracy
  • Magnus Vinding
  • Eva Vivalt
  • Jackson Wagner
  • Ben West | Philosophy for Programmers
  • Nick Whitaker | High Modernism
  • Jess Whittlestone
  • Peter Wildeford | Everyday Utilitarian
  • Peter Wildeford | Not Quite a Blog
  • Peter Wildeford | Pasteur’s Cube
  • Peter Wildeford | The Power Law
  • Bridget Williams
  • Ben Williamson
  • Julia Wise | Giving Gladly
  • Julia Wise | Otherwise
  • Julia Wise | The Whole Sky
  • Kat Woods
  • Thomas Woodside
  • Mark Xu | Artificially Intelligent
  • Linch Zhang
  • Anisha Zaveri

NEWSLETTERS

  • AI Safety Newsletter
  • Alignment Newsletter
  • Altruismo eficaz
  • Animal Ask’s Newsletter
  • Apart Research
  • Astera Institute | Human Readable
  • Campbell Collaboration Newsletter
  • CAS Newsletter
  • ChinAI Newsletter
  • CSaP Newsletter
  • EA Forum Digest
  • EA Groups Newsletter
  • EA & LW Forum Weekly Summaries
  • EA UK Newsletter
  • EACN Newsletter
  • Effective Altruism Lowdown
  • Effective Altruism Newsletter
  • Effective Altruism Switzerland
  • Epoch Newsletter
  • Farm Animal Welfare Newsletter
  • Existential Risk Observatory Newsletter
  • Forecasting Newsletter
  • Forethought
  • Future Matters
  • GCR Policy’s Newsletter
  • Gwern Newsletter
  • High Impact Policy Review
  • Impactful Animal Advocacy Community Newsletter
  • Import AI
  • IPA’s weekly links
  • Manifold Markets | Above the Fold
  • Manifund | The Fox Says
  • Matt’s Thoughts In Between
  • Metaculus – Medium
  • ML Safety Newsletter
  • Open Philanthropy Farm Animal Welfare Newsletter
  • Oxford Internet Institute
  • Palladium Magazine Newsletter
  • Policy.ai
  • Predictions
  • Rational Newsletter
  • Sentinel
  • SimonM’s Newsletter
  • The Anti-Apocalyptus Newsletter
  • The EA Behavioral Science Newsletter
  • The EU AI Act Newsletter
  • The EuropeanAI Newsletter
  • The Long-termist’s Field Guide
  • The Works in Progress Newsletter
  • This week in security
  • TLDR AI
  • Tlön newsletter
  • To Be Decided Newsletter

PODCASTS

  • 80,000 Hours Podcast
  • Alignment Newsletter Podcast
  • AI X-risk Research Podcast
  • Altruismo Eficaz
  • CGD Podcast
  • Conversations from the Pale Blue Dot
  • Doing Good Better
  • Effective Altruism Forum Podcast
  • ForeCast
  • Found in the Struce
  • Future of Life Institute Podcast
  • Future Perfect Podcast
  • GiveWell Podcast
  • Gutes Einfach Tun
  • Hear This Idea
  • Invincible Wellbeing
  • Joe Carlsmith Audio
  • Morality Is Hard
  • Open Model Project
  • Press the Button
  • Rationally Speaking Podcast
  • The End of the World
  • The Filan Cabinet
  • The Foresight Institute Podcast
  • The Giving What We Can Podcast
  • The Good Pod
  • The Inside View
  • The Life You Can Save
  • The Rhys Show
  • The Turing Test
  • Two Persons Three Reasons
  • Un equilibrio inadecuado
  • Utilitarian Podcast
  • Wildness

VIDEOS

  • 80,000 Hours: Cambridge
  • A Happier World
  • AI In Context
  • AI Safety Talks
  • Altruismo Eficaz
  • Altruísmo Eficaz
  • Altruísmo Eficaz Brasil
  • Brian Tomasik
  • Carrick Flynn for Oregon
  • Center for AI Safety
  • Center for Security and Emerging Technology
  • Centre for Effective Altruism
  • Center on Long-Term Risk
  • Chana Messinger
  • Deliberation Under Ideal Conditions
  • EA for Christians
  • Effective Altruism at UT Austin
  • Effective Altruism Global
  • Future of Humanity Institute
  • Future of Life Institute
  • Giving What We Can
  • Global Priorities Institute
  • Harvard Effective Altruism
  • Leveller
  • Machine Intelligence Research Institute
  • Metaculus
  • Ozzie Gooen
  • Qualia Research Institute
  • Raising for Effective Giving
  • Rational Animations
  • Robert Miles
  • Rootclaim
  • School of Thinking
  • The Flares
  • EA Groups
  • EA Tweets
  • EA Videos
Inactive websites grayed out.
Effective Altruism News is a side project of Tlön.
  • Important Insights
    30 pillars of my worldview
    Bentham's Newsletter | 1 hours ago
  • A Version of Direct Realism
    Here, I explain the sense in which sensory perception provides direct awareness of the external world. *
    Fake Nous | 2 hours ago
  • Drugs Aren't A Moral Category
    Are drugs good?. This question doesn't really make sense. Yet Western society answers with a firm "NO". I have ADHD and have a prescription for Methylphenidate (MPH). Often I don't feel like taking it. Shouldn't I be able to just do the right things? I can just decide to be productive. Right? Needing a drug to function feels like a flaw in who I am. I also have sleep apnea.
    LessWrong | 2 hours ago
  • Vaccine Obstructionism Kills
    The FDA kills far more people than vaccines do
    Good Thoughts | 2 hours ago
  • What makes a good test of AI intelligence?
    #ai #aialignment #cognitivescience...
    Rational Animations | 3 hours ago
  • Legitimate Deliberation
    Legitimate deliberation is an alignment target; an alternative to / variation of Coherent Extrapolated Volition. What could make us trust an AI? Can we imagine a near-future scenario where we might consider some frontier model's outputs to be reliable, rather than needing to check any claims it makes via human sources?
    LessWrong | 3 hours ago
  • The global decline in murder, explained in one chart
    One source of good news — favored both by me and, apparently, venture capitalists — is what’s known as a “narrative violation.” A narrative violation occurs when everyone thinks one thing, but the actual evidence suggests the opposite. And few narratives are more persistently violated than one common belief: “Violent crime is always going up.” A 2023 […]...
    Future Perfect | 4 hours ago
  • The environment is a terrible reason to avoid ChatGPT
    People are saying you shouldn’t use ChatGPT due to statistics like:
    Benjamin Todd | 4 hours ago
  • A Playbook for New EA Groups
    A note from Naomi Nederlof, Community Building Grants Program Manager: I asked Andy, the lead organizer of EA DC, to write this playbook based on the approaches that have made his group one of the most successful city groups in our program. We initiated this project because we think many new organizers can benefit from concrete and experience-based guidance.
    Effective Altruism Forum | 6 hours ago
  • Collective intelligence, under the influence
    From the annals of “Research it must have been fun to do” - getting people drunk to study the benefits group decision-making.
    Reasonable People | 10 hours ago
  • The Missing Genre: Heroic Parenthood - You can have kids and still punch the sun
    I stopped reading when I was 30. You can fill in all the stereotypes of a girl with a book glued to her face during every meal, every break, and 10 hours a day on holidays. That was me. And then it was not. For 9 years I’ve been trying to figure out why. I mean, I still read. Technically. But not with the feral devotion from Before. And I finally figured out why.
    LessWrong | 10 hours ago
  • Should you work with evil people?
    Epistemic status: Figuring things out. My mind often wanders to what boundaries I ought to maintain between the different parts of my life and people who have variously committed bad acts or have poor character. On the professional side, I think it is a virtue to be able to work with lots of people, and be functional in many environments.
    LessWrong | 11 hours ago
  • Does climate science focus on the right temperature range?
    As long as we are not in the year 2100, we will not know exactly what path humanity will have taken when it comes to climate change in this century. But we can at least make educated guesses about the trajectory we are on. Given this, it seems that a good amount of climate science is aimed at hopeful temperature ranges below 2°C, which we will likely miss.
    Effective Altruism Forum | 12 hours ago
  • Why Wouldn't A Rationalist Be Rational?
    Sometimes people ask me why someone behaving irrationally is present at a rationalist meetup. This essay is a partial answer to why I think that question is confused. The short version is, we are not set up such that any particular list of skills is a requirement.
    LessWrong | 12 hours ago
  • Food Fridays: Thanksgiving
    I have decided to do Food Fridays on at least some Fridays.
    Thing of Things | 15 hours ago
  • Tests of LLM introspection need to rule out causal bypassing
    This point has been floating around implicitly in various papers (e.g., Betley et al., Plunkett et al., Lindsey), but we haven’t seen it named explicitly. We think it’s important, so we’re describing it here. There’s been growing interest in testing whether LLMs can introspect on their internal states or processes.
    LessWrong | 17 hours ago
  • Save Big and Make a Difference: Mercy For Animals’ Black Friday Sale
    Get ready to boost your holiday cheer with Mercy For Animals’ Black Friday bonanza! Now through November 30, enjoy savings of 15% or more on everything from our store. When you shop our selection of Mercy For Animals logo apparel, drinkware, holiday cards, and gifts, you’re doing more than snagging great deals—you’re actively supporting our […].
    Mercy for Animals | 19 hours ago
  • Ruby's Ultimate Guide to Thoughtful Gifts
    Give a man a gift and he smiles for a day. Teach a man to gift and he’ll cause smiles for the rest of his life. Gift giving is an exercise in theory of mind, empathy, noticing, and creativity. “What I discovered is that my girlfiend wants me to give her gifts the way you give gifts.” – words from a friend. .
    LessWrong | 19 hours ago
  • A Thanksgiving Memory
    A couple of days before Thanksgiving when I was 11 years old, the house my family lived in was gutted by a fire. We had been stripping the white paint off the 150-year-old oak woodwork as a restoration project. It was snowing and windy out, so the workers had closed the windows. A spark from plugging in a work lamp lit the turpentine fumes, and a fireball tore through the house.
    LessWrong | 19 hours ago
  • A Taxonomy of Bugs (Lists)
    One of my favorite basic concepts from CFAR is the Bugs List. By writing down everything in your life that feels "off," things you'd change if you could, problems you've been meaning to solve, irritations you've learned to live with, you have a straightforward set of problems to try solving with new techniques and frameworks you learn, and can, little by little, actually improve your life in a...
    LessWrong | 19 hours ago
  • You Are Much More Salient To Yourself Than To Everyone Else
    Back in the Boy Scouts, at summer camp, myself and a couple friends snuck out one night after curfew to commandeer a couch someone had left by a dumpster at the other end of the camp (maybe a half kilometer away). Now, our particular designated adult was a very stick-to-the-rules type, so we definitely did not want to get caught. I, therefore, made my way slowly and sneakily.
    LessWrong | 19 hours ago
  • The Best Lack All Conviction: A Confusing Day in the AI Village
    The AI Village is an ongoing experiment (currently running on weekdays from 10 a.m. to 2 p.m. Pacific time) in which frontier language models are given virtual desktop computers and asked to accomplish goals together. Since Day 230 of the Village (17 November 2025), the agents' goal has been "Start a Substack and join the blogosphere".
    LessWrong | 19 hours ago
  • Claude Opus 4.5: Model Card, Alignment and Safety
    They saved the best for last. The contrast in model cards is stark. Google provided a brief overview of its tests for Gemini 3 Pro, with a lot of ‘we did this test, and we learned a lot from it, and we are not going to tell you the results.’. Anthropic gives us a 150 page book, including their capability assessments. This makes sense.
    LessWrong | 19 hours ago
  • Writing advice: Why people like your quick bullshit takes better than your high-effort posts
    Right now I’m coaching for Inkhaven, a month-long marathon writing event where our brave residents are writing a blog post every single day for the entire month of November. And I’m pleased that some of them have seen success – relevant figures seeing the posts, shares on Hacker News and Twitter and LessWrong.
    LessWrong | 21 hours ago
  • Why people like your quick bullshit takes better than your high-effort posts
    Right now I’m coaching for Inkhaven, a month-long marathon writing event where our brave residents are writing a blog post every single day for the entire month of November. And I’m pleased that some of them have seen success – relevant figures seeing the posts, shares on Hacker News and Twitter and LessWrong.
    LessWrong | 21 hours ago
  • Why people like your quick bullshit takes better than your high-effort posts
    Those new bloggers, who are kind of upset about the internet's bad taste, might benefit from reading artist Dimespin’s essay: “Why people like your doodles better than your finished works.”
    Eukaryote Writes Blog | 21 hours ago
  • Donation Election Discussion Thread
    This is a thread for explaining your vote, discussing it, and maybe changing your mind. It'll be pinned on the frontpage throughout the Donation Election. Some comments on this thread are cross-posted from a text box which appears when you reach the end of the voting process, but everyone is welcome to post here whenever.
    Effective Altruism Forum | 1 days ago
  • Eliezer Yudkowsky's View On Consciousness Doesn't Make Any Sense
    You shouldn't do neuroscience from the armchair, especially if it leads you to conclude that toddlers aren't conscious
    Bentham's Newsletter | 1 days ago
  • Will AI safety become a mass movement?
    Some AI safety activists think the community should borrow from the climate playbook and build broad public appeal — but not everyone agrees...
    Transformer | 1 days ago
  • How Accurate Are Audits In Assessing Animal Welfare?
    The consumer push for better treatment of farmed animals has led to a growth in supply chain audits. But are current audits accurate enough? This study looks at the reliability of one of these tools in evaluating the welfare of pigs on U.S. farms. The post How Accurate Are Audits In Assessing Animal Welfare? appeared first on Faunalytics.
    Faunalytics | 1 days ago
  • A brief guide to the groups protesting over AI
    The differences between Stop AI, PauseAI, ControlAI, and more
    Transformer | 1 days ago
  • The goal of AI companies is to replace people
    Future of Life Institute | 1 days ago
  • The AI safety checklist (we’re failing every single requirement)
    Future of Life Institute | 1 days ago
  • Despair, Serenity, Song and Nobility in "Hollow Knight: Silksong"
    Fictional universes are oft defined by what their positive affect feels like and what their negative affect feels like. This is the palette that the story is written using. In the videogame Silksong, the negative affect is a hollow, pointless, loneliness.
    LessWrong | 1 days ago
  • Consciousness and Competition (with Joe Carlsmith)
    Joe Carlsmith is a writer, researcher, and philosopher. He works at Anthropic on the character/constitution/spec for Claude. Before that, he was a senior advisor at Open Philanthropy. To see all Forethought’s published research, visit forethought.org/research. To subscribe to our newsletter, visit forethought.org/subscribe.
    ForeCast | 1 days ago
  • Outcome robustness and implementation robustness
    Anthony DiGiovanni Credit to Anni Leskelä for coming up with this framework. We’d like to take interventions that look net-positive even after accounting for our deep uncertainty about the long-term future. I’ll say an intervention is robust to the extent that the argument for the intervention being net-positive is not sensitive to unusually hard-to-estimate factors.
    Center on Long-Term Risk | 1 days ago
  • Wartime ethics is weird
    The ethical principles that most people hold—and hold most strongly—go completely out the window when it comes to war. Normal time: Killing is bad. In fact it’s pretty much the worst thing you can do. Wartime: Killing is great! Kill as many people as you can! If you’re really good at killing, you get a medal!. (Just so long as you kill the right people.).
    Philosophical Multicore | 1 days ago
  • Will We Get Alignment by Default? — with Adrià Garriga-Alonso
    Adrià recently published “Alignment will happen by default; what’s next?” on LessWrong, arguing that AI alignment is turning out easier than expected. Simon left a lengthy comment pushing back, and that sparked this spontaneous debate. Adrià argues that current models like Claude Opus 3 are genuinely good “to their core,” and that an iterative process — where each AI generation helps align...
    LessWrong | 1 days ago
  • Information Hygiene
    Do avalanches get caused by loud noises?. Based on my dozen+ times giving this class or presentation, at least 7/10 of you are nodding yes, and the main reason the other 3 aren’t is that you sense a trap. So. What do you think you know, and why do you think you know it?. Our bodies are under constant assault.
    LessWrong | 1 days ago
  • Underrated reasons to be thankful: AI research edition
    I experienced my first Thanksgiving and thus had to write this.
    AI Safety Takes | 2 days ago
  • Hidden Open Thread 409.5
    Astral Codex Ten | 2 days ago
  • iPhone Fold leaks 📱, lightning on Mars ⚡, technical deflation 📉
    TLDR AI | 2 days ago
  • ALLFED's 2025 Highlights
    Executive Summary. Welcome to ALLFED’s 2025 Highlights, our annual look back at how Alliance to Feed the Earth in Disasters has advanced its mission to help build resilience to global catastrophic food system failure. Within These 2025 Highlights:
    Effective Altruism Forum | 2 days ago
  • A Technical Introduction to Solomonoff Induction without K-Complexity
    This post has been written in collaboration with Iliad in service of one of Iliad's longer-term goals of understanding the simplicity bias of learning machines. Solomonoff induction is a general theoretical ideal for how to predict sequences that were generated by a computable process.
    LessWrong | 2 days ago
  • Enjoy the Silence
    This holiday season, AI companies are working tirelessly with your lawmakers to make sure *every* night is silent.
    Future of Life Institute | 2 days ago
  • Dark secrets of the alpha male revealed
    I have, over the course of my life, had the opportunity to observe many men who have sex with lots and lots of women.
    Thing of Things | 2 days ago
  • Highlights from Ilya Sutskever’s November 2025 interview with Dwarkesh Patel
    Background. Ilya Sutskever is a renowned machine learning researcher who co-authored the AlexNet paper in 2012 that helped kick off the deep learning revolution. Sutskever co-founded OpenAI and served as Chief Scientist until 2024. He participated in the temporary ouster of CEO Sam Altman and subsequently left the company.
    Effective Altruism Forum | 2 days ago
  • Liking Clear Writing Isn't A Fetish, Actually!
    Responding to professor Ellie Anderson's reply to me on continental philosophy
    Bentham's Newsletter | 2 days ago
  • Help revive EA at UPenn (looking for board members)
    The UPenn Effective Altruism group has been inactive since spring 2024. A small group and I are now reviving it!. I'm Hazem, a freshman in Mechanical Engineering at UPenn and one of two current organizers. Current status: 2 organizers, 2 advisors, ~10 interested non-organizing members. 3 socials held. access restored to club infrastructure (website, logins, PennClubs.com, etc.).
    Effective Altruism Forum | 2 days ago
  • Billionaires are building a future without humans
    AI companies are explicitly working toward AGI and are likely to succeed soon, possibly within years. Keep the Future Human explains how unchecked development of smarter-than-human, autonomous, general-purpose AI systems will almost inevitably lead to human replacement. But it doesn't have to.
    Future of Life Institute | 2 days ago
  • The Case For A European Union-Wide “Positive List”
    A positive list, which focuses on species that can be safely and humanely kept as companion animals, might just be the solution the European Union needs to address the harms of the exotic pet trade. The post The Case For A European Union-Wide “Positive List” appeared first on Faunalytics.
    Faunalytics | 2 days ago
  • Malengo Impact Model Update, Fall 2025
    Johannes Haushofer and Richard Nerland. Summary: Malengo is a charity that facilitates international educational migration in the pursuit of poverty alleviation. Its flagship program supports Ugandan high school graduates in applying for and completing a Bachelor’s degree in Germany.
    Effective Altruism Forum | 2 days ago
  • Cómo hacer tu plan de carrera profesional
    Unirse a una comunidad es una estrategia muy eficaz para avanzar en la carrera profesional y amplificar el impacto positivo de cada uno, superando los beneficios de las redes de contactos convencionales. El valor proviene de la dinámica colaborativa que permite a los grupos lograr más que la suma de los esfuerzos individuales de sus miembros a través de la especialización, los recursos...
    Altruismo Eficaz | 2 days ago
  • ASTMH 2025: Advancing Gene Drive Conversations
    The American Society of Tropical Medicine and Hygiene (ASTMH) Annual Meeting is the world’s leading forum for tropical medicine and global health. This year, for the first time, it was held outside the U.S., in Toronto, Canada, bringing together thousands of experts to share research and foster collaboration. The meeting featured vibrant discussions on malaria, […].
    Target Malaria | 2 days ago
  • A twice-yearly vitamin A capsule costs just over $1 and can save a child’s sight and life.
    Giving What We Can | 2 days ago
  • if you don't get enough you go blind
    A single capsule of vitamin A given to children twice a year for the first five years of their lives can save their sight and lives, for only a little more than $1 per dose
    Giving What We Can | 2 days ago
  • Why OpenAI Is Trying to Silence Its Critics (with Tyler Johnston)
    Tyler Johnston is Executive Director of the Midas Project. He joins the podcast to discuss AI transparency and accountability. We explore applying animal rights watchdog tactics to AI companies, the OpenAI Files investigation, and OpenAI's subpoenas against nonprofit critics.
    Future of Life Institute | 2 days ago
  • Announcing our Latest Influenced-Giving Metrics
    See how Animal Charity Evaluators is impacting the animal advocacy movement. …  Read more...
    Animal Charity Evaluators | 2 days ago
  • Meet CRS Event Recording
    CRS hosted a public end-of-year call to share more about what CRS is working on, how we approach reducing suffering, and what we see as the key priorities ahead. The event included a short presentation on our current work, followed by an open Q&A with CRS founders Magnus Vinding and Tobias Baumann. Thank you to […].
    Center for Reducing Suffering | 2 days ago
  • Alignment remains a hard, unsolved problem
    This is a public adaptation of a document I wrote for an internal Anthropic audience about a month ago. Thanks to (in alphabetical order) Joshua Batson, Joe Benton, Sam Bowman, Roger Grosse, Jeremy Hadfield, Jared Kaplan, Jan Leike, Jack Lindsey, Monte MacDiarmid, Sam Marks, Fra Mosconi, Chris Olah, Ethan Perez, Sara Price, Ansh Radhakrishnan, Fabien Roger, Buck Shlegeris, Drake Thomas, and...
    LessWrong | 2 days ago
  • Alignment remains a hard, unsolved problem
    This is a public adaptation of a document I wrote for an internal Anthropic audience about a month ago. Thanks to (in alphabetical order) Joshua Batson, Joe Benton, Sam Bowman, Roger Grosse, Jeremy Hadfield, Jared Kaplan, Jan Leike, Jack Lindsey, Monte MacDiarmid, Sam Marks, Fra Mosconi, Chris Olah, Ethan Perez, Sara Price, Ansh Radhakrishnan, Fabien Roger, Buck Shlegeris, Drake Thomas, and...
    AI Alignment Forum | 2 days ago
  • Alignment Bootstrapping Is Dangerous
    AI companies want to bootstrap weakly-superhuman AI to align superintelligent AI. I don’t expect them to succeed. I could give various arguments for why alignment bootstrapping is hard and why AI companies are ignoring the hard parts of the problem; but you don’t need to understand any details to know that it’s a bad plan.
    Philosophical Multicore | 2 days ago
  • Even superhuman AI forecasters are only as good as your questions
    Predicting the future is difficult. It might become easier. What then?
    AI Safety Takes | 2 days ago
  • To write better, just explain it to someone
    I was once helping a child with her homework. She was supposed to write about a place that was important for her, and had chosen her family’s summer cottage. Every now and then, she would get distracted from the writing, tell me something about the cottage, and then complain that she didn’t know what to write next. Me: “Well, you could write the thing that you just told me.”. Child: “Oh!
    LessWrong | 3 days ago
  • Subliminal Learning Across Models
    Tl;dr: We show that subliminal learning can transfer sentiment across models (with some caveats). For example, we transfer positive sentiment for Catholicism, the UK, New York City, Stalin or Ronald Reagan across model families using normal-looking text. This post discusses under what conditions this subliminal transfer happens. —.
    LessWrong | 3 days ago
  • Against The Omnipresent Advantage Argument For Trans Sports
    Astral Codex Ten | 3 days ago
  • iPhone dominance 📈, S&P downgrades Tether 🪙, lessons on agents 🤖
    TLDR AI | 3 days ago
  • 🎁 ¿Qué regalar estas Navidades?
    Ayuda Efectiva | 3 days ago
  • Principles and Generators of a Rationality Dojo
    Since 2023 I've been directing WARP, the Wandering Applied Rationality Program, with the help of ESPR and SPARC staff, which are summer camps I've taught at since 2017. For those that don't know, these ~10 day programs are hard to summarize well, but are generally meant to create an environment that helps participants better understand the world, themselves, and each other by offering a...
    LessWrong | 3 days ago
  • Microsoft’s Fairwater datacenter will use more power than Los Angeles
    The next generation of AI data center campuses are city-scale.
    Epoch Newsletter | 3 days ago
  • Postmodernism for STEM Types: A Clear-Language Guide to Conflict Theory
    Crossposted from Susbstack. Section I : Opening. In 2021, Richard Dawkins tweeted: . The fallout was immediate. The American Humanist Association revoked an award they’d given him 25 years earlier. A significant controversy erupted, splitting roughly into two camps. One camp defended Dawkins. They saw him raising a legitimate question about logical consistency.
    LessWrong | 3 days ago
  • Forecasting Claude
    Above the Fold is "Playing it Safe"
    Manifold Markets | 3 days ago
  • Training PhD Students to be Fat Newts (Part 2)
    [Thanks Inkhaven for hosting me! This is my fourth and last post and I'm already exhausted from writing. Wordpress.com!]. Last time, I introduced the concept of the “Fat Newt” (fatigue neutral) build, a way of skilling up characters in Battle Brothers that aims to be extremely economical with the fatigue resource, relying entirely on each brother’s base 15 fatigue regeneration per turn.
    LessWrong | 3 days ago
  • Snippets on Living In Reality
    Social reality is quite literally another world, in the same sense that the Harry Potter universe is another world. Like the Harry Potter universe, social reality is a world portrayed primarily in text and in speech and in our imaginations. Like the Harry Potter universe, social reality doesn’t diverge completely from physical reality - they contain mostly the same cities, for instance.
    LessWrong | 3 days ago
  • My idea of what's, like, #3 trending on Redtube right now
    Content warnings: sex, dubcon, drug use.
    Thing of Things | 3 days ago
  • Courtship Confusions Post-Slutcon
    Going into slutcon, one of my main known-unknowns was… I’d heard many times that the standard path to hooking up or dating starts with two people bantering for an hour or two at a party, lacing in increasingly-unsubtle hints of interest. And even in my own imagination, I was completely unable to make up a plausible-sounding conversation which would have that effect.
    LessWrong | 3 days ago
  • AI discourse analyzed (we looked at essays, Twitter, Bluesky, Truth Social)
    AI In Group Discourse. I wanted to programmatically analyze the AI in-group ecosystem and discourse using AI as an exploration in sensemaking during my time in the “ AI for Human Reasoning” FLF fellowship.
    Effective Altruism Forum | 3 days ago
  • Training PhD Students to be Fat Newts (Part 1)
    Today, I want to introduce an experimental PhD student training philosophy. Let’s start with some reddit memes. Every gaming subreddit has its own distinct meme culture. On r/chess, there's a demon who is summoned by an unsuspecting beginner asking “Why isn’t this checkmate?”. . These posts are gleefully deluged by responses saying “Google En Passant” in some form or other.
    LessWrong | 3 days ago
  • The Economics of Replacing Call Center Workers With AIs
    TLDR: Voice AIs aren't that much cheaper in the year 2025: My friend runs a voice agent startup in Canada for walk-in clinics. The AI takes calls and uses tools to book appointments in the EMR (electronic medical record) system. In theory, this helps the clinic hire less front desk staff and the startup makes infinite money. In reality, the margins are brutal and they barely charge above cost.
    LessWrong | 3 days ago
  • Evaluating honesty and lie detection techniques on a diverse suite of dishonest models
    TL;DR: We use a suite of testbed settings where models lie—i.e. generate statements they believe to be false—to evaluate honesty and lie detection techniques. The best techniques we studied involved fine-tuning on generic anti-deception data and using prompts that encourage honesty. Read the full Anthropic Alignment Science blog post and the X thread. Introduction:
    LessWrong | 3 days ago
  • Evolution & Freedom
    In Against Money Maximalism, I argued against money-maximization as a normative stance. Profit is a coherent thing you can try to maximize, but there are also other sorts of value. Profit-maximization isn't the unique rational way to engage with money. One way you could respond to this is economic Darwinism: "Sure, you can optimize other things than money, but over time, the market will come...
    LessWrong | 3 days ago
  • Subliminal Learning Across Models
    Tl;dr: We show that subliminal learning can transfer sentiment across models (with some caveats). For example, we transfer positive sentiment for Catholicism, the UK, New York City, Stalin or Ronald Reagan across model families using normal-looking text. This post discusses under what conditions this subliminal transfer happens. —.
    AI Alignment Forum | 3 days ago
  • Donation Election, diversity talent bank, career paths in Africa
    Your farmed animal advocacy update for late November 2025
    Hive | 3 days ago
  • Building the Fortress
    Reframing Suffering-Focused Ethics By David Veldran Many moral views and social projects present themselves as inherently positive and constructive. They aim to add something to the world, to create, to build. Classical utilitarians seek to increase happiness and bring about a surplus of joy over misery. Communists aspire to realize a classless society. Kantians, perhaps, aim to […].
    Center for Reducing Suffering | 3 days ago
  • Scared to Pledge? 5 financial steps for confident giving
    Cross-posted from the @High Impact Professionals blog - see the original here. --. In my years of working with people on their financial plans for giving, I’ve met many wonderful individuals who feel morally compelled to take a giving pledge. Intellectually, they are fully aware that they are in a position to help others. But emotionally, they aren’t ready to commit to a pledge.
    Effective Altruism Forum | 3 days ago
  • The history of vaccines
    The early smallpox vaccines that kept dying out, why Émile Roux drilled into rabbits' skulls, and the lucky career changes that saved millions of lives.
    The Works in Progress Newsletter | 3 days ago
  • You don't want to date multiple people who are monogamous with you
    Sometimes people say, “obviously, everyone would like to be polyamorous themselves and get to have sex with anyone they want, while their partners are all monogamous and only have sex with them.
    Thing of Things | 3 days ago
  • Public Acceptability Of Standard U.S. Animal Agriculture Practices
    A Faunalytics study reveals that a clear majority of the U.S. public finds standard animal agriculture practices for pigs, cows, and chickens to be unacceptable. Explore the details and their implications for animal advocacy. The post Public Acceptability Of Standard U.S. Animal Agriculture Practices appeared first on Faunalytics.
    Faunalytics | 3 days ago
  • SB 53 protects whistleblowers in AI — but asks a lot in return
    Opinion: Abra Ganz and Karl Koch argue that whistleblower protections in SB-53 aren’t good enough on the face of it — but how the state chooses to interpret the law could turn that around...
    Transformer | 3 days ago
  • For A Short Period Of Time, You Can Save 21,000 Shrimp Per Dollar
    5 arguments for doing so!
    Bentham's Newsletter | 3 days ago
  • Four Key Ways To Help Farmed Animals In Nigeria
    Although public awareness of animal welfare is relatively low in Nigeria, this report provides a roadmap for advocates looking to make a difference for the country’s growing number of farmed animals. The post Four Key Ways To Help Farmed Animals In Nigeria appeared first on Faunalytics.
    Faunalytics | 3 days ago
  • Why the West was downzoned
    In the space of a few decades, nearly every city in the Western world banned densification. What happened?
    The Works in Progress Newsletter | 3 days ago
  • Better for Animals: The Evidence Behind Protests
    Animal Charity Evaluators’ Better for Animals project summarizes research on animal advocacy interventions. Our third spotlight is a review of the evidence behind protests. …  Read more...
    Animal Charity Evaluators | 3 days ago
  • 2025 Results, 2026 Plans and Funding Needs
    We have just released our end-of-year 2025 Results, 2026 Plans, and Funding Needs document that demonstrates what a donation to RP can accomplish: from contributing to better allocation of millions of…...
    Rethink Priorities | 3 days ago
  • Is This Anything? 22
    Recently I've been getting into Gratitude Patrols. People talk about gratitude journals as being one of the few provably impactful psychological interventions [citation needed], but I think there's a benefit to physically embodying it by walking around your space each morning and thanking things. If you.
    Atoms vs Bits | 3 days ago
  • Nutrition Investing: Moving from Awareness to Action
    Nutrition Investing: Moving from Awareness to Action gloireri Wed, 11/26/2025 - 14:04 . Investing in nutrition isn’t just possible, it’s smart. That’s the key message that sticks with us a few weeks after the GIIN Impact Forum 2025, where we organised a session, “Nutrition Lens Investing: A Framework for Action”.
    Global Alliance for Improved Nutrition | 3 days ago
  • 3 doubts about veganism
    I keep thinking about what kind of identity would be useful for building a powerful animal advocacy movement. Here are 3 features of veganism that I often think about which make me doubt its usefulness. Too maximalist. The official definition of veganism by the inventors of the term is the following:
    Effective Altruism Forum | 3 days ago
  • Understanding Moral Disagreement Through Data
    We studied moral judgments in the U.S. and found extreme disagreement. Learn how 15 moral dimensions predict people’s moral views.
    Clearer Thinking | 3 days ago
  • Nuevo ranking de «Top invitadores» | ¿Qué regalar estas Navidades?
    Ayuda Efectiva | 3 days ago
  • EA Forum Digest #268
    EA Forum Digest #268 Hello!. Three things to note: The Donation Election is now open! You can go vote on the Forum (if you had an account before October 24th). Read about all the candidates here (if you spot an error in this post, I have to donate $10), and read the rules here.
    EA Forum Digest | 3 days ago

Loading...

SEARCH

SUBSCRIBE

RSS Feed
X Feed

ORGANIZATIONS

  • Effective Altruism Forum
  • 80,000 Hours
  • Ada Lovelace Institute
  • Against Malaria Foundation
  • AI Alignment Forum
  • AI Futures Project
  • AI Impacts
  • AI Now Institute
  • AI Objectives Institute
  • AI Safety Camp
  • AI Safety Communications Centre
  • AidGrade
  • Albert Schweitzer Foundation
  • Aligned AI
  • ALLFED
  • Asterisk
  • altLabs
  • Ambitious Impact
  • Anima International
  • Animal Advocacy Africa
  • Animal Advocacy Careers
  • Animal Charity Evaluators
  • Animal Ethics
  • Apollo Academic Surveys
  • Aquatic Life Institute
  • Association for Long Term Existence and Resilience
  • Ayuda Efectiva
  • Berkeley Existential Risk Initiative
  • Bill & Melinda Gates Foundation
  • Bipartisan Commission on Biodefense
  • California YIMBY
  • Cambridge Existential Risks Initiative
  • Carnegie Corporation of New York
  • Center for Applied Rationality
  • Center for Election Science
  • Center for Emerging Risk Research
  • Center for Health Security
  • Center for Human-Compatible AI
  • Center for Long-Term Cybersecurity
  • Center for Open Science
  • Center for Reducing Suffering
  • Center for Security and Emerging Technology
  • Center for Space Governance
  • Center on Long-Term Risk
  • Centre for Effective Altruism
  • Centre for Enabling EA Learning and Research
  • Centre for the Governance of AI
  • Centre for the Study of Existential Risk
  • Centre of Excellence for Development Impact and Learning
  • Charity Entrepreneurship
  • Charity Science
  • Clearer Thinking
  • Compassion in World Farming
  • Convergence Analysis
  • Crustacean Compassion
  • Deep Mind
  • Democracy Defense Fund
  • Democracy Fund
  • Development Media International
  • EA Funds
  • Effective Altruism Cambridge
  • Effective altruism for Christians
  • Effective altruism for Jews
  • Effective Altruism Foundation
  • Effective Altruism UNSW
  • Effective Giving
  • Effective Institutions Project
  • Effective Self-Help
  • Effective Thesis
  • Effektiv-Spenden.org
  • Eleos AI
  • Eon V Labs
  • Epoch Blog
  • Equalize Health
  • Evidence Action
  • Family Empowerment Media
  • Faunalytics
  • Farmed Animal Funders
  • FAST | Animal Advocacy Forum
  • Felicifia
  • Fish Welfare Initiative
  • Fistula Foundation
  • Food Fortification Initiative
  • Foresight Institute
  • Forethought
  • Foundational Research Institute
  • Founders’ Pledge
  • Fortify Health
  • Fund for Alignment Research
  • Future Generations Commissioner for Wales
  • Future of Life Institute
  • Future of Humanity Institute
  • Future Perfect
  • GBS Switzerland
  • Georgetown University Initiative on Innovation, Development and Evaluation
  • GiveDirectly
  • GiveWell
  • Giving Green
  • Giving What We Can
  • Global Alliance for Improved Nutrition
  • Global Catastrophic Risk Institute
  • Global Challenges Foundation
  • Global Innovation Fund
  • Global Priorities Institute
  • Global Priorities Project
  • Global Zero
  • Good Food Institute
  • Good Judgment Inc
  • Good Technology Project
  • Good Ventures
  • Happier Lives Institute
  • Harvard College Effective Altruism
  • Healthier Hens
  • Helen Keller INTL
  • High Impact Athletes
  • HistPhil
  • Humane Slaughter Association
  • IDInsight
  • Impactful Government Careers
  • Innovations for Poverty Action
  • Institute for AI Policy and Strategy
  • Institute for Progress
  • International Initiative for Impact Evaluation
  • Invincible Wellbeing
  • Iodine Global Network
  • J-PAL
  • Jewish Effective Giving Initiative
  • Lead Exposure Elimination Project
  • Legal Priorities Project
  • LessWrong
  • Let’s Fund
  • Leverhulme Centre for the Future of Intelligence
  • Living Goods
  • Long Now Foundation
  • Machine Intelligence Research Institute
  • Malaria Consortium
  • Manifold Markets
  • Median Group
  • Mercy for Animals
  • Metaculus
  • Metaculus | News
  • METR
  • Mila
  • New Harvest
  • Nonlinear
  • Nuclear Threat Initiative
  • One Acre Fund
  • One for the World
  • OpenAI
  • Open Mined
  • Open Philanthropy
  • Organisation for the Prevention of Intense Suffering
  • Ought
  • Our World in Data
  • Oxford Prioritisation Project
  • Parallel Forecast
  • Ploughshares Fund
  • Precision Development
  • Probably Good
  • Pugwash Conferences on Science and World Affairs
  • Qualia Research Institute
  • Raising for Effective Giving
  • Redwood Research
  • Rethink Charity
  • Rethink Priorities
  • Riesgos Catastróficos Globales
  • Sanku – Project Healthy Children
  • Schmidt Futures
  • Sentience Institute
  • Sentience Politics
  • Seva Foundation
  • Sightsavers
  • Simon Institute for Longterm Governance
  • SoGive
  • Space Futures Initiative
  • Stanford Existential Risk Initiative
  • Swift Centre
  • The END Fund
  • The Future Society
  • The Life You Can Save
  • The Roots of Progress
  • Target Malaria
  • Training for Good
  • UK Office for AI
  • Unlimit Health
  • Utility Farm
  • Vegan Outreach
  • Venten | AI Safety for Latam
  • Village Enterprise
  • Waitlist Zero
  • War Prevention Initiative
  • Wave
  • Wellcome Trust
  • Wild Animal Initiative
  • Wild-Animal Suffering Research
  • Works in Progress

PEOPLE

  • Scott Aaronson | Shtetl-Optimized
  • Tom Adamczewski | Fragile Credences
  • Matthew Adelstein | Bentham's Newsletter
  • Matthew Adelstein | Controlled Opposition
  • Vaidehi Agarwalla | Vaidehi's Blog
  • James Aitchison | Philosophy and Ideas
  • Scott Alexander | Astral Codex Ten
  • Scott Alexander | Slate Star Codex
  • Scott Alexander | Slate Star Scratchpad
  • Alexanian & Franz | GCBR Organization Updates
  • Applied Divinity Studies
  • Leopold Aschenbrenner | For Our Posterity
  • Amanda Askell
  • Amanda Askell's Blog
  • Amanda Askell’s Substack
  • Atoms vs Bits
  • Connor Axiotes | Rules of the Game
  • Sofia Balderson | Hive
  • Mark Bao
  • Boaz Barak | Windows On Theory
  • Nathan Barnard | The Good blog
  • Matthew Barnett
  • Ollie Base | Base Rates
  • Simon Bazelon | Out of the Ordinary
  • Tobias Baumann | Cause Prioritization Research
  • Tobias Baumann | Reducing Risks of Future Suffering
  • Nora Belrose
  • Rob Bensinger | Nothing Is Mere
  • Alexander Berger | Marginal Change
  • Aaron Bergman | Aaron's Blog
  • Satvik Beri | Ars Arcana
  • Aveek Bhattacharya | Social Problems Are Like Maths
  • Michael Bitton | A Nice Place to Live
  • Liv Boeree
  • Dillon Bowen
  • Topher Brennan
  • Ozy Brennan | Thing of Things
  • Catherine Brewer | Catherine’s Blog
  • Stijn Bruers | The Rational Ethicist
  • Vitalik Buterin
  • Lynette Bye | EA Coaching
  • Ryan Carey
  • Joe Carlsmith
  • Lucius Caviola | Outpaced
  • Richard Yetter Chappell | Good Thoughts
  • Richard Yetter Chappell | Philosophy, Et Cetera
  • Paul Christiano | AI Alignment
  • Paul Christiano | Ordinary Ideas
  • Paul Christiano | Rational Altruist
  • Paul Christiano | Sideways View
  • Paul Christiano & Katja Grace | The Impact Purchase
  • Evelyn Ciara | Sunyshore
  • Cirrostratus Whispers
  • Jesse Clifton | Jesse’s Substack
  • Peter McCluskey | Bayesian Investor
  • Greg Colbourn | Greg's Substack
  • Ajeya Cotra & Kelsey Piper | Planned Obsolescence
  • Owen Cotton-Barrat | Strange Cities
  • Andrew Critch
  • Paul Crowley | Minds Aren’t Magic
  • Dale | Effective Differentials
  • Max Dalton | Custodienda
  • Saloni Dattani | Scientific Discovery
  • Amber Dawn | Contemplatonist
  • De novo
  • Michael Dello-Iacovo
  • Jai Dhyani | ANOIEAEIB
  • Michael Dickens | Philosophical Multicore
  • Dieleman & Zeijlmans | Effective Environmentalism
  • Anthony DiGiovanni | Ataraxia
  • Kate Donovan | Gruntled & Hinged
  • Dawn Drescher | Impartial Priorities
  • Eric Drexler | AI Prospects: Toward Global Goal Convergence
  • Holly Elmore
  • Sam Enright | The Fitzwilliam
  • Daniel Eth | Thinking of Utils
  • Daniel Filan
  • Lukas Finnveden
  • Diana Fleischman | Dianaverse
  • Diana Fleischman | Dissentient
  • Julia Galef
  • Ben Garfinkel | The Best that Can Happen
  • Matthew Gentzel | The Consequentialist
  • Aaron Gertler | Alpha Gamma
  • Rachel Glennerster
  • Sam Glover | Samstack
  • Andrés Gómez Emilsson | Qualia Computing
  • Nathan Goodman & Yusuf | Longterm Liberalism
  • Ozzie Gooen | The QURI Medley
  • Ozzie Gooen
  • Katja Grace | Meteuphoric
  • Katja Grace | World Spirit Sock Puppet
  • Katja Grace | Worldly Positions
  • Spencer Greenberg | Optimize Everything
  • Milan Griffes | Flight from Perfection
  • Simon Grimm
  • Zach Groff | The Groff Spot
  • Erich Grunewald
  • Marc Gunther
  • Cate Hall | Useful Fictions
  • Chris Hallquist | The Uncredible Hallq
  • Topher Hallquist
  • John Halstead
  • Finn Hambly
  • James Harris | But Can They Suffer?
  • Riley Harris
  • Riley Harris | Million Year View
  • Peter Hartree
  • Shakeel Hashim | Transformer
  • Sarah Hastings-Woodhouse
  • Hayden | Ethical Haydonism
  • Julian Hazell | Julian’s Blog
  • Julian Hazell | Secret Third Thing
  • Jiwoon Hwang
  • Incremental Updates
  • Roxanne Heston | Fire and Pulse
  • Hauke Hillebrandt | Hauke’s Blog
  • Ben Hoffman | Compass Rose
  • Michael Huemer | Fake Nous
  • Tyler John | Secular Mornings
  • Mike Johnson | Open Theory
  • Toby Jolly | Seeking To Be Jolly
  • Holden Karnofsky | Cold Takes
  • Jeff Kaufman
  • Cullen O'Keefe | Jural Networks
  • Daniel Kestenholz
  • Ketchup Duck
  • Oliver Kim | Global Developments
  • Isaac King | Outside the Asylum
  • Petra Kosonen | Tiny Points of Vast Value
  • Victoria Krakovna
  • Ben Kuhn
  • Alex Lawsen | Speculative Decoding
  • Gavin Leech | argmin gravitas
  • Howie Lempel
  • Gregory Lewis
  • Eli Lifland | Foxy Scout
  • Rhys Lindmark
  • Robert Long | Experience Machines
  • Garrison Lovely | Garrison's Substack
  • MacAskill, Chappell & Meissner | Utilitarianism
  • Jack Malde | The Ethical Economist
  • Jonathan Mann | Abstraction
  • Sydney Martin | Birthday Challenge
  • Daniel May
  • Daniel May
  • Conor McCammon | Utopianish
  • Peter McIntyre
  • Peter McIntyre | Conceptually
  • Pablo Melchor
  • Pablo Melchor | Altruismo racional
  • Geoffrey Miller
  • Fin Moorhouse
  • Luke Muehlhauser
  • Neel Nanda
  • David Nash | Global Development & Economic Advancement
  • David Nash | Monthly Overload of Effective Altruism
  • Eric Neyman | Unexpected Values
  • Richard Ngo | Narrative Ark
  • Richard Ngo | Thinking Complete
  • Elizabeth Van Nostrand | Aceso Under Glass
  • Oesterheld, Treutlein & Kokotajlo | The Universe from an Intentional Stance
  • James Ozden | Understanding Social Change
  • Daniel Paleka | AI Safety Takes
  • Ives Parr | Parrhesia
  • Dwarkesh Patel | The Lunar Society
  • Kelsey Piper | The Unit of Caring
  • Michael Plant | Planting Happiness
  • Michal Pokorný | Agenty Dragon
  • Georgia Ray | Eukaryote Writes Blog
  • Ross Rheingans-Yoo | Icosian Reflections
  • Josh Richards | Tofu Ramble
  • Jess Riedel | foreXiv
  • Anna Riedl
  • Hannah Ritchie | Sustainability by Numbers
  • David Roodman
  • Eli Rose
  • Abraham Rowe | Good Structures
  • Siebe Rozendal
  • John Salvatier
  • Anders Sandberg | Andart II
  • William Saunders
  • Joey Savoie | Measured Life
  • Stefan Schubert
  • Stefan Schubert | Philosophical Sketches
  • Stefan Schubert | Stefan’s Substack
  • Nuño Sempere | Measure is unceasing
  • Harish Sethu | Counting Animals
  • Rohin Shah
  • Zeke Sherman | Bashi-Bazuk
  • Buck Shlegeris
  • Jay Shooster | jayforjustice
  • Carl Shulman | Reflective Disequilibrium
  • Jonah Sinick
  • Andrew Snyder-Beattie | Defenses in Depth
  • Ben Snodin
  • Nate Soares | Minding Our Way
  • Kaj Sotala
  • Tom Stafford | Reasonable People
  • Pablo Stafforini | Pablo’s Miscellany
  • Henry Stanley
  • Jacob Steinhardt | Bounded Regret
  • Zach Stein-Perlman | AI Lab Watch
  • Romeo Stevens | Neurotic Gradient Descent
  • Michael Story | Too long to tweet
  • Maxwell Tabarrok | Maximum Progress
  • Max Taylor | AI for Animals
  • Benjamin Todd
  • Brian Tomasik | Reducing Suffering
  • Helen Toner | Rising Tide
  • Philip Trammell
  • Jacob Trefethen
  • Toby Tremlett | Raising Dust
  • Alex Turner | The Pond
  • Alyssa Vance | Rationalist Conspiracy
  • Magnus Vinding
  • Eva Vivalt
  • Jackson Wagner
  • Ben West | Philosophy for Programmers
  • Nick Whitaker | High Modernism
  • Jess Whittlestone
  • Peter Wildeford | Everyday Utilitarian
  • Peter Wildeford | Not Quite a Blog
  • Peter Wildeford | Pasteur’s Cube
  • Peter Wildeford | The Power Law
  • Bridget Williams
  • Ben Williamson
  • Julia Wise | Giving Gladly
  • Julia Wise | Otherwise
  • Julia Wise | The Whole Sky
  • Kat Woods
  • Thomas Woodside
  • Mark Xu | Artificially Intelligent
  • Linch Zhang
  • Anisha Zaveri

NEWSLETTERS

  • AI Safety Newsletter
  • Alignment Newsletter
  • Altruismo eficaz
  • Animal Ask’s Newsletter
  • Apart Research
  • Astera Institute | Human Readable
  • Campbell Collaboration Newsletter
  • CAS Newsletter
  • ChinAI Newsletter
  • CSaP Newsletter
  • EA Forum Digest
  • EA Groups Newsletter
  • EA & LW Forum Weekly Summaries
  • EA UK Newsletter
  • EACN Newsletter
  • Effective Altruism Lowdown
  • Effective Altruism Newsletter
  • Effective Altruism Switzerland
  • Epoch Newsletter
  • Farm Animal Welfare Newsletter
  • Existential Risk Observatory Newsletter
  • Forecasting Newsletter
  • Forethought
  • Future Matters
  • GCR Policy’s Newsletter
  • Gwern Newsletter
  • High Impact Policy Review
  • Impactful Animal Advocacy Community Newsletter
  • Import AI
  • IPA’s weekly links
  • Manifold Markets | Above the Fold
  • Manifund | The Fox Says
  • Matt’s Thoughts In Between
  • Metaculus – Medium
  • ML Safety Newsletter
  • Open Philanthropy Farm Animal Welfare Newsletter
  • Oxford Internet Institute
  • Palladium Magazine Newsletter
  • Policy.ai
  • Predictions
  • Rational Newsletter
  • Sentinel
  • SimonM’s Newsletter
  • The Anti-Apocalyptus Newsletter
  • The EA Behavioral Science Newsletter
  • The EU AI Act Newsletter
  • The EuropeanAI Newsletter
  • The Long-termist’s Field Guide
  • The Works in Progress Newsletter
  • This week in security
  • TLDR AI
  • Tlön newsletter
  • To Be Decided Newsletter

PODCASTS

  • 80,000 Hours Podcast
  • Alignment Newsletter Podcast
  • AI X-risk Research Podcast
  • Altruismo Eficaz
  • CGD Podcast
  • Conversations from the Pale Blue Dot
  • Doing Good Better
  • Effective Altruism Forum Podcast
  • ForeCast
  • Found in the Struce
  • Future of Life Institute Podcast
  • Future Perfect Podcast
  • GiveWell Podcast
  • Gutes Einfach Tun
  • Hear This Idea
  • Invincible Wellbeing
  • Joe Carlsmith Audio
  • Morality Is Hard
  • Open Model Project
  • Press the Button
  • Rationally Speaking Podcast
  • The End of the World
  • The Filan Cabinet
  • The Foresight Institute Podcast
  • The Giving What We Can Podcast
  • The Good Pod
  • The Inside View
  • The Life You Can Save
  • The Rhys Show
  • The Turing Test
  • Two Persons Three Reasons
  • Un equilibrio inadecuado
  • Utilitarian Podcast
  • Wildness

VIDEOS

  • 80,000 Hours: Cambridge
  • A Happier World
  • AI In Context
  • AI Safety Talks
  • Altruismo Eficaz
  • Altruísmo Eficaz
  • Altruísmo Eficaz Brasil
  • Brian Tomasik
  • Carrick Flynn for Oregon
  • Center for AI Safety
  • Center for Security and Emerging Technology
  • Centre for Effective Altruism
  • Center on Long-Term Risk
  • Chana Messinger
  • Deliberation Under Ideal Conditions
  • EA for Christians
  • Effective Altruism at UT Austin
  • Effective Altruism Global
  • Future of Humanity Institute
  • Future of Life Institute
  • Giving What We Can
  • Global Priorities Institute
  • Harvard Effective Altruism
  • Leveller
  • Machine Intelligence Research Institute
  • Metaculus
  • Ozzie Gooen
  • Qualia Research Institute
  • Raising for Effective Giving
  • Rational Animations
  • Robert Miles
  • Rootclaim
  • School of Thinking
  • The Flares
  • EA Groups
  • EA Tweets
  • EA Videos
Inactive websites grayed out.
Effective Altruism News is a side project of Tlön.