![]() "Minimum Viable Coup" is my new favorite concept. From Dwarkesh interviewing Paul Christiano, asking "what's the minimum capabilities needed for a superintelligent AI to overthrow the government?" Read more: All
0 Comments
If they're not conscious, we still have to worry about instrumental convergence. Viruses are dangerous even if they're not conscious. But if they are conscious, we have to worry that we are monstrous slaveholders causing Black Mirror nightmares for the sake of drafting emails to sell widgets. Of course, they might not care about being turned off. But there's already empirical evidence of them spontaneously developing self-preservation goals (because you can't achieve your goals if you're turned off). Read more: All
He had a falling out with one abolitionist leader and faction, who then spent time and money spreading rumors about him and posting flyers around each town in his lecture circuit, calling him a fraud. Usually this was over what in retrospect seems really trivial things, and surely they could have still worked together or at least peacefully pursue separate strategies (e.g. should they prioritize legal reform or changing public opinion? Did one activist cheat on his wife with a colleague?) Reading his biography, it's unclear who attacked him more: the slave owners or his fellow abolitionists. In-fighting is part of every single movement I’ve ever read about. EA and AI safety are not special in that regard. “I am not at all surprised when some of those for whom I have lived and labored lift their heels against me. Since the days of Moses such has been the fate of all men earnestly endeavouring to serve the oppressed and unfortunate.”
It doesn’t mean internet mobs aren’t also terrible to deal with, but it reminds me to feel grateful for our current state. If you do advocacy nowadays, you must fear character assassination, but rarely physical assassination (at least in democratic rich countries).
Quote from the book where some other abolitionists thought he was bad for the movement because he wasn’t arguing about obscure Constitutional law and was instead trying to appeal to a larger audience with vaguer messages. Reminds me of the debates over AI safety comms, where some people want things to be precise and dry and maximally credible to academics, and other people want to appeal to a larger audience using emotions, metaphor, and not getting into arcane details
Emphasizes that humor is a way to spread your message. People are more likely to listen if you mix in laugher with getting them to look at the darkness.
He was a leader, and he knew that without hope, people wouldn’t fight.
He was ahead of the curve on women’s rights, which is no small feat in the 1800s. But he was also a temperance advocate, being against alcohol. And he really hated Catholics. It’s a good reminder to be humble about your ethical beliefs. If you spend a lot of time thinking about ethics and putting it into practice, you’ll likely be ahead of your time in some ways. But you’ll also probably be wrong about some things. Remember - the road to hell isn’t paved with good intentions. It’s paved with overconfident intentions.
Moral suasion is a persuasive technique that uses rhetorical appeals and persuasion to change a person or group's behavior. It's a non-coercive way to influence people to act in a certain way.
Loved this excerpt: Treated as a “deserter from the fold,” he nevertheless, or so he claimed, let his colleagues “search me and probe me to the bottom.” Facing what he considered outright lies, he stood firm against the hailstorm of “side blows, innuendo, dark suspicions, such as avarice, faithlessness, treachery, ingratitude and what not.” Whistling in the graveyard, he assured Smith proudly that he felt “strengthened to bear it without perturbation.” And this line: “Turning affliction into hope, however many friends he might lose“
“I would unite with anybody to do right,” he said, “and with nobody to do wrong.” “I contend that I have a right to cooperate with anybody with everybody for the overthrow of slavery” “Stop seeking purity, he told his critics among radicals, and start with what is possible”
He cheated on his wife. He was racist (against the Irish and Native Americans), prejudiced against Catholics, and overly sensitive to perceived slights. And yet, he is a moral hero nevertheless. Don’t expect perfection from anybody, including yourself. Practice the virtues of understanding and forgiveness, and we’re all better off.
Not a lesson learned really, but had to be said. Seriously, the book has a gorgeous cover, has the cool roughcut edges of the pages, has a properly serious looking “Winner of Pullitzer Prize” award on the front, feels just the right level of heavy, and is just the most satisfying weighty tome. Referring to the hardcover edition of David W Blight’s biography. Read more: All We had a whole class of people for ages who had nothing to do but hangout with people and attend parties. Just read any Jane Austen novel to get a sense of what it's like to live in a world with no jobs. Only a small fraction of people, given complete freedom from jobs, went on to do science or create something big and important. Most people just want to lounge about and play games, watch plays, and attend parties. They are not filled with angst around not having a job. In fact, they consider a job to be a gross and terrible thing that you only do if you must, and then, usually, you must minimize. Our society has just conditioned us to think that jobs are a source of meaning and importance because, well, for one thing, it makes us happier. We have to work, so it's better for our mental health to think it's somehow good for us. And for two, we need money for survival, and so jobs do indeed make us happier by bringing in money. Massive job loss from AI will not by default lead to us leading Jane Austen lives of leisure, but more like Great Depression lives of destitution. We are not immune to that. Us having enough is incredibly recent and rare, historically and globally speaking. Remember that approximately 1 in 4 people don't have access to something as basic as clean drinking water. You are not special. You could become one of those people. You could not have enough to eat. So AIs causing mass unemployment is indeed quite bad. But it's because it will cause mass poverty and civil unrest. Not because it will cause a lack of meaning. (Of course I'm more worried about extinction risk and s-risks. But I am more than capable of worrying about multiple things at once) Read more: All “The difference between nuclear arms treaties and AI treaties is that it’s so easy to copy AIs, so regulation is hopeless” This is only true for existing models. Inventing new, state of the art models is incredibly difficult and expensive. It requires immense amounts of talent, infrastructure, money, compute, and innovations that people don’t yet know how to do. Almost all of the human extinction risk from AIs come from not-yet-invented superintelligent AI models. North Korea or a terrorist group cannot just defect from an AI treaty and build superintelligent AI. And it’s relatively straightforward to monitor and prevent the amount of compute necessary to make a superintelligent AI (e.g. monitoring electrical grids, specialized GPUs, satellite imagery, etc) Once it’s already invented, then yes, people could easily steal it. But if we just stop sometime 𝘣𝘦𝘧𝘰𝘳𝘦 we have superintelligent AI, then it will be very hard for any group to defect. Also, by the time we superintelligent AI, it’s probably already too late, and it will be up to the superintelligence what to do, not humans anymore. Read more: All Do you live in Berkeley? Are you feeling particularly anxious or depressed? Consider moving to SF or elsewhere. I've heard from so many people that the AI safety community in Berkeley is particularly damaging to mental health. I've only felt depressed about AI safety once & it was when I was in Berkeley for a few weeks. It went away after I left. From what I've heard it seems to be from a mix of:
Remember: you're the average of the 5 people you spend the most time with. You might be depressed or anxious because of objective facts about the world. But you might also be depressed or anxious because you're surrounded by depressed or anxious people I've heard good things about the SF AI safety community. London's EA community was also far less neurotic in my experience than many others. Felt much more balanced and less insular. Read more: All Disclaimer: this will only work for a subset of you. Law of Equal and Opposite Advice and all that. It might only even work for me. This definitely feels like a weird psychological trick that might only work with my brain. I spent my twenties being absolutely devastated by uncertainty. I saw the suffering in the world and I desperately wanted to help, but the more I learned and the more I tried, the wider my confidence intervals got. Maybe I could promote bednets. But what about the meat eater problem? Maybe I could promote veganism? But what about the small animal replacement problem? Even giving out free hugs (the most clearly benign thing I could think of) might cause unexpected trauma for some unknown percentage of the population such that it negates all the positives. It eventually reached a crescendo in 2020 where I sunk into absolute epistemic hopelessness. An RCT had just been published about the intervention I was doing that didn't even show that the intervention didn't work. It was just ambiguous. If at least it had been obviously zero impact, I could have moved on. But it was ambiguous for goodness sake! I actually briefly gave up on altruism. I was going to go be a hippie in the woods and make art and do drugs. After all, if I couldn't know if what I was doing was helping or even hurting, I might as well be happy myself. But then…. I saw something in the news about the suffering in the world. And I wanted to help. No, a part of me said. You can't help, remember? Nothing works. Or you can never tell if it's working. And then another thing showed up in my social media feed…. But no! It wasn’t worth trying because the universe was too complex and I was but a monkey in shoes. But still. . . . another part of me couldn’t look away. It said “Look at the suffering. You can’t possibly see that and not at least try.” I realized in that moment that I couldn’t actually be happy if I wasn’t at least trying. This led to a large breakthrough in how I felt. Before, there was always the possibility of stopping and just having fun. So I was comparing all of the hard work and sacrifice I was doing to this ideal alternative life. When I realized that even if I had basically no hope, I’d still keep trying, this liberated me. There was no alternative life where I wasn’t trying. It felt like the equivalent of burning the ships. No way to go but forward. No temptation of retreat. Many things aren’t bad in and of themselves, but bad compared to something else. If you remove the comparison, then they’re good again. But it wasn’t over yet. I was still deeply uncertain. I went to Rwanda to try to actually get as close to ground truth as possible, while also reading a ton about meta-ethics, to get at the highest level stuff, then covid hit. While I was stuck in lockdown, I realized that I should take the simulation hypothesis seriously. You’d think this would intensify my epistemic nihilism, but it didn’t. It turned me into an epistemic absurdist. Which is basically the same thing, but happy. Even if this is base reality, I’m profoundly uncertain about whether bednets are even net positive. Now you add that this might all be a simulation?!? For real?! (Pun was unintentional but appreciated, so I’m keeping it) This was a blessing in disguise though, because suddenly it went from:
The more certain you feel, the more you feel you can control things, and that leads to feeling more stressed out. As you become more uncertain, it can feel more and more stressful, because there’s an outcome you care about and you’re not sure how to get there. But if you have only very minimal control, you can either freak out more, because it’s out of your control, or you can relax, because it’s out of your control. So I became like the Taoist proverb: "A drunkard falls out of a carriage but doesn't get hurt because they go limp." If somebody walked by a drowning child that would be trivially easy to save, I’d think they were a monster. If somebody walks by a deeply complex situation where getting involved may or may not help and may even accidentally make it worse, but then tries to help anyway, I think they’re a good person and if it doesn’t work out, well, hey, at least they tried. I relaxed into the uncertainty. The uncertainty means I don’t have to be so hard on myself, because it’s just too complicated to really know one way or the other. Nowadays I work in AI safety, and whenever I start feeling anxious about timelines and p(doom), the most reliable way for me to feel better is to remind myself about the deep uncertainty around everything. “Remember, this might all be a simulation. And even if it isn’t, it’s really hard to figure out what’s net positive, so just do something that seems likely to be good, and make sure it’s something you at least enjoy, so no matter what, you’ll at least have had a good life” How can other people apply this? I think this won’t work for most people, but you can try this on and see if it works for you:
Anyways, while I’m sure this won’t work for most people, hopefully some people who are currently struggling in epistemic nihilism might be able to come out the other side and enjoy epistemic absurdism like me. But in the end, who knows? Read more: All You can't just say “epistemic status: garbage” and then hope your words don’t have consequences1/25/2025 You can't just say “epistemic status: garbage” and then hope your words don’t have consequences Especially online, where everything you write is forever and can be damaging a person’s reputation, mental health, and ability to do good until the singularity. This isn’t to say you shouldn’t criticise. But you should publicly criticize when you have decent evidence and arguments, not when it’s just hearsay, vibe, anonymous sources, or any other form of terrible epistemics. Read more: All
Read more: All 𝐍𝐞𝐰 𝐦𝐞𝐧𝐭𝐚𝐥 𝐡𝐞𝐚𝐥𝐭𝐡 𝐩𝐫𝐨𝐠𝐫𝐚𝐦 𝐟𝐨𝐫 𝐩𝐞𝐨𝐩𝐥𝐞 𝐰𝐨𝐫𝐤𝐢𝐧𝐠 𝐨𝐧 𝐀𝐈 𝐬𝐚𝐟𝐞𝐭𝐲! 🥳 It’s not therapy. It’s what I wish therapy was, but totally isn’t. It’s a short program that lasts 4-12 weeks, where you systematically try 5-30 techniques until you find something that fixes an emotional problem you're struggling with (e.g. anxiety, impostor syndrome, low mood, etc). Here’s how it works: 𝐅𝐢𝐫𝐬𝐭 𝐜𝐚𝐥𝐥: 𝐮𝐧𝐝𝐞𝐫𝐬𝐭𝐚𝐧𝐝 𝐚𝐧𝐝 𝐩𝐥𝐚𝐧
You’ll spend the next 1-3 weeks actually putting the most promising techniques into practice. You’ll keep track of your symptoms. If your symptoms go away, then we’ll analyze what happened. Sometimes it’ll be obvious what’s helping, and you can just keep doing that thing. If not, then we can start remove the techniques one at a time. If the symptoms come back, then we just bring back the technique that we removed, and we know what was doing the magic. Experimenting in parallel means you get to feel better sooner and continue to feel good while we figure out what the problem was. If your symptoms don’t go away after 1-2 weeks, then we’ll prioritize the next 5-10 techniques to try. This process will happen up to 3 times. By the end, you’ll have either resolved your issues, or you’ll at least have tried ~30 techniques to fix the problem. Even if you haven’t, you’ll probably have found at least a few more techniques to add to your repertoire of things that you enjoy. Apply here 𝐈𝐬 𝐭𝐡𝐢𝐬 𝐭𝐡𝐞𝐫𝐚𝐩𝐲? It’s not therapy. It’s what I wish therapy was, but totally isn’t. 𝐄𝐦𝐨𝐭𝐢𝐨𝐧𝐚𝐥 𝐩𝐫𝐨𝐛𝐥𝐞𝐦𝐬 𝐭𝐡𝐞 𝐩𝐫𝐨𝐠𝐫𝐚𝐦 𝐜𝐚𝐧 𝐡𝐞𝐥𝐩 𝐰𝐢𝐭𝐡: Stress Impostor syndrome Burnout Anxiety Hopelessness Feeling overwhelmed Depression (mild or moderate. Not severe) Self-esteem issues Motivation issues Numbness Sadness Work life balance Guilt Sleep issues Loneliness Existential angst Perfectionism Relationship problems 𝐄𝐦𝐨𝐭𝐢𝐨𝐧𝐚𝐥 𝐩𝐫𝐨𝐛𝐥𝐞𝐦𝐬 𝐭𝐡𝐞 𝐩𝐫𝐨𝐠𝐫𝐚𝐦 𝙘𝙖𝙣𝙣𝙤𝙩 𝐡𝐞𝐥𝐩 𝐰𝐢𝐭𝐡 Suicidality Bipolar ADHD Gender dysphoria Anger management Substance use disorders Autism related emotional issues Cluster B personality disorders (e.g. BPD, APD, HPD, NPD) Anything where you're experiencing psychosis Anything where you're experiencing paranoia or delusions 𝐇𝐨𝐰 𝐦𝐮𝐜𝐡 𝐝𝐨𝐞𝐬 𝐢𝐭 𝐜𝐨𝐬𝐭? It's free if you:
I'm offering this service for free because mental health is one of the main blockers to people having an impact in AI safety. I think x-risks and s-risks from AI are the most important things to work on. And I'm good at emotional problem-solving. So if I help people working on AI safety be happier, then I'm helping make sure AI doesn't kill everybody. Timelines are too short to work with people who are not working in AI safety, but since I'm a rationalist and everybody has their price, I would do it for a non-AI safety person if they donated $10,000 or more to Nonlinear or an AI safety org working on pausing or slowing down AI development. Apply here I have very limited time, so can only take on a small fraction of clients who apply. If you do not get in, I recommend checking out this vetted list of therapists or this compilation of mental health techniques for dealing with AI safety. Read more: All |
Popular postsThe Parable of the Boy Who Cried 5% Chance of Wolf
The most important lesson I learned after ten years in EA Why fun writing can save lives Full List Categories
All
Kat WoodsI'm an effective altruist who co-founded Nonlinear, Charity Entrepreneurship, and Charity Science Health Archives
February 2025
Categories
All
|