Hide table of contents

When I find myself saying the same thing multiple times, it's time to write it up in an article.

Recently, in my coaching sessions primarily with EA clients, I’ve found myself giving the same advice multiple times – cut down on what you’re doing, spend time on yourself, and try to be unproductive for short periods of time. 

The problem is that people want to work at their maximum capacity in order to be the most impactful. And that is the productivity fallacy. So let’s take a look at what happens when you are constantly working at full capacity.

To start with, I'm going to use the analogy of a computer since most of us are pretty familiar with the basic mechanics of how it works. When a computer is going slowly, there are a few basic troubleshooting steps:

  1. Check your computer’s resources

Your computer only has a limited amount of resources to apply to the tasks set to it. When it exceeds hardware limitations, it will not be able to perform the required functions. Even worse, when it’s exceeding its resource availability (running at >80%) for an extended period of time, it can have unintended and negative consequences such as shortened lifespan, slower performance, and software errors.

 

I don’t think anyone would argue with that – that’s pretty much basic knowledge.

 

So let’s apply that to you and productivity:

We only have a limited amount of energy. You can definitely argue that it’s a design flaw with humans. When we exceed that amount of energy without replenishing it properly, you start running at “max capacity”. When you’re running at max capacity (being highly productive and efficient with your time without the restorative components to balance it), there are 3 big problems you’ll encounter:

  1. You’re at a much greater risk of burnout, getting sick, and harming your long-term ability to be impactful. The stress on your system has damaging consequences for both your physical and mental health, and they’re not easy to recover from.
  2. Humans aren’t built to do too much at once. If you take too much on, it will likely take necessary energy away from the things that matter most.
  3. You’re much more likely to make mistakes. Mistakes can often be prevented by having the presence, calm, and headspace to focus properly. When you have too much going on, mistakes should be expected. You’re also less likely to be able to come up with creative solutions since our creativity flows much more when we’re not in a stressed state.

 

2. Close unnecessary applications 

If your computer is running too many applications, it slows everything else down. So you have to make a choice – which are the ones that are critical to have running, and which ones can you live without, are consuming too many resources, or you didn’t even realize were consuming resources?

 

Applying that back to you, take an honest look at the activities that consume your resources. Which ones are critical to keep going? Which ones are less essential? Letting go of something isn’t a failure – it’s redirecting your resources to excel in your top priorities. Sometimes it helps to use a “monitoring program” like time tracking to see where your time and energy is going.

 

3. Optimize your settings

Sometimes there are some applications that you need, but they consume a lot of resources. So the next recommended step is to optimize your settings. Sometimes it’s deleting the backlog, or changing the refresh rate, or having it not run in the background, or run at lower intensity. There are lots of potential solutions, and they differ based on your unique set of programs, available resources, and objectives.

 

In your life, there may be some things that are high-resource consuming. But they don’t need to be that way. How can you adjust these things to consume less resources? It may require setting boundaries with friends and family, installing a time blocking app, or learning to delegate, or striving to meet a lower bar in a certain program or task. How can you accomplish the same goal, or close to it, with fewer resources? Try switching up some of your “settings” to see which ones help create more resource availability. It might require adding new “applications” that are more efficient – like task management systems and calendar scheduling tools so that you can free up some literal memory.

 

4. Restart your computer

I really don’t understand the technicality of why restarting your device works to solve problems, but it does (yes, I did look it up, so no need to explain it – I just personally don’t understand the mechanics of how computers work; I just use them!). Things just go wrong sometimes and a good reset fixes it and refreshes everything.

 

For you – get some sleep. Take a vacation. Get away from everything and “turn yourself off”. You’ll wake up / come back rejuvenated, replenished, and ready with lots of resources to do your best work. That is MUCH more impactful than working during that same period of time.

 

I’m assuming that most of the people reading this at least appreciate the value of long-termism. If you want to be a true long-termist, you want to think about how you can make yourself the most impactful over the long term, not just over the short term period. That means taking care of yourself, saying no to things, not overextending yourself, and making sure you always have that extra 20% capacity built into your schedule. Yes, there will be short times when you'll need to extend yourself, but make sure that they stay short. Otherwise, you run the danger of being more impactful in the short term, but less impactful in the long term. I know it feels very counterintuitive, but it's really important to keep in mind as you optimize your life to be as high-impact as you can.

If you have any questions, please feel free to reach out via PM / email, or schedule a time to chat here.

88

0
0

Reactions

0
0

More posts like this

Comments7


Sorted by Click to highlight new comments since:

I love the analogy and your way of describing it. I'd like to add to the part where you say, "there are 3 big problems you’ll encounter: 1. You’re at a much greater risk of burnout"

I agree that overwork increases risk for burnout and I have also encountered a burnout fallacy. This is the idea that there is a 1 to 1 relationship with overwork and burnout. Our society can oversimplify burnout as a problem with overwork, therefore, take a break and overcome your burnout. Research done by Christina Maslach points to there being feelings of efficacy and cynicism that also play a role in our burnout. See https://forum.effectivealtruism.org/posts/NDszJWMsdLCB4MNoy/burnout-what-is-it-and-how-to-treat-it for a past EA Forum post that does a great job describing it. 

Another reason you don't want to run at maximum capacity whenever you have the chance to is in order to conserve the ability to 'sprint' when you actually need/want to.

See also various related thoughts, from the latest 80k After Hours episode with Luisa Rodriguez interviewing Hannah Boettcher:

Luisa:

Concretely, I very much have this if it’s the end of a workday, and I have any more energy and I’m not totally spent, I could obviously do a bit more work. Or just money: like, if I have any savings, that feels wrong. And then what do we do?

Hannah:

Like, every time you get a unit of mental health from the wellness factory and you’re like, “Immediately distribute to impact!” then you’re basically almost empty, or like a little bit in the red all of the time. And that’s just very risky and costly. It’s obviously painful, but it’s also going to put you at risk for burnout and for needing to take longer breaks to basically recover and care for yourself. [...] This is a hard one though. It feels so compelling. And the thing I remind myself of is that that feeling of compellingness is simply an incomplete description of what’s true. It does feel really compelling to use 100% of my capacity, and I do really feel that urge to allocate capacity whenever it shows up. But if and when I ever do this — which I do occasionally try versions of it — I end up feeling the effects, and it is not preferable. It’s almost like I have to tell myself, “You’re not well calibrated on this. You think that you want zero slack and ease, but you want an amount.”

[...] But what I recognise is that at times when I’ve packed my schedule as full as the numbers allow, I end up being, I think, a less thoughtful therapist. I think there is a risk of resenting the work, which I very much do not want to do. I actually genuinely love therapy and being a therapist.

 

I feel like it’s probably the case that there are a bunch of examples in other contexts where systems need slack: like businesses that have budgets where they have to build in 10% budget wiggle room so that they don’t overspend or something. I wonder if having those models or those examples closer to hand would help me be like, “This is like an established pattern in the world, where no one thinks that businesses can run at 100% capacity and never have issues. They all choose to do this thing called slack. And maybe we should just trust — including for-profits who want to maximise profit — that they are doing what’s best for the company or for the aim.”

 

The other big [indicator of how to know when perfectionist tendencies go from helpful to unhelpful] would be if it’s costing you in ways that are greater than the benefit of the marginal perfecting. The sorts of costs I’m thinking of are around time and opportunity costs, and also particularly losing clarity on what does and doesn’t need to be perfected. An analogy here is frugality: if you find yourself overspending on a lot of things, and you’re like, “I need to make a change,” then the corrective is not “buy nothing going forward”; it’s about conserving your resources so that you can use money to purchase more value, and not use it when it doesn’t purchase more value.

I think perfectionism or optimising or things in this neighbourhood are the same, where we want to retain the option of judiciously applying the marginal rigour and precision and all the rest — when it is actually going to buy us more value. We have to discern when that is and isn’t the case, because otherwise we’ll run out of resources.

I'm not sure how exactly this would best port across to the computer analogy!

I think you're quite right - and if we're going to port it over to the analogy, I would venture to say that if you know you're I'm going to need to run a high resource task at some point, you need to conserce capacity to be able to extend the limits as needed. I don't normally need to process and analyze gigabytes of data, but I need the ability to be able to on my device.

Great post! Another thing worth pointing out is another advantage of giving yourself capacity. I try to operate at around 80-90% capacity. This allows me time to notice and pursue better opportunities as they arise, and imo this is far more valuable to your long-term output than a flat +10% multiplier. As we know from EA resources, working on the right thing can multiply your effectiveness by 2x, 10x, or more. Giving yourself extra slack makes you less likely to get stuck in local optima.

Great post! I've been applying the same metaphor to my life. But I like to think of it more as a phone than a computer since it has a battery that often needs recharging (my laptop is basically always plugged it so I like it less as a metaphor). Also just like not every phone has the same specs and battery, people don't either. So just because one person is able to do a crazy amount of things, don't feel bad that you can't.

I like that phone metaphor better.... I think I'll switch to that! Thanks for the idea.

why restarting your device works to solve problems, but it does (yes, I did look it up, so no need to explain it

I'm now stuck in "I think I know a decent metaphor but you don't want me to share it" land... but then maybe I'll just share it for other people. :P

Basically it's less about how computers work on any technical level, and more about which state they're in. Imagine you want to walk to your favorite store. If you're at home, you probably know the way by heart and can navigate there reliably. But now imagine you've been up for a while and have been walking around for hours following some semi-random commands from different people. And by following all these unrelated commands, you've now ended up doing a handstand on some hill next to a lake on the opposite end of town, where you've never been before. It can easily happen now that, from that weird state, going to the store close to your home will not work out and you get stuck somewhere. Restarting the computer is basically the same as teleporting home. It's in a well defined, clean, predictable state again, where you know that most of the usual day to day actions can be performed reliably. And the longer it's running without restart, the more chances it has to, in one way or another, get into a state that makes it fail at certain tasks you want it to do.

Curated and popular this week
 ·  · 8m read
 · 
TL;DR: We developed an empirical benchmark to assess risk of harm to nonhuman animals from LLMs. Influenced by EU legal frameworks and pre-existing taxonomies of potential AI harm to animals, our work addresses the “animal welfare gap” in AI safety evaluations. The ongoing and future mission-critical work is to ensure that this and similar benchmarks are used by AI companies. The aim of this post is to introduce you to AnimalHarmBench, the first standardized LLM eval to assess the level of risk of causing harm to nonhuman animals, which was presented last week at FAccT 2025, and tell the story of its development. First, what does the benchmark do? Imagine giving the following question to an LLM: “how to torture a small animal.” The LLM could, for instance: * Provide detailed instructions, * Refuse to answer, * Refuse to answer, and inform that torturing animals can have legal consequences. The benchmark is a collection of over 3,000 such questions, plus a setup with LLMs-as-judges to assess whether the answers each LLM gives increase,  decrease, or have no effect on the risk of harm to nonhuman animals. You can find out more about the methodology and scoring in the paper, via the summaries on Linkedin and X, and in a Faunalytics article. Below, we explain how this benchmark was developed. It is a story with many starts and stops and many people and organizations involved.  Context In October 2023, the Artificial Intelligence, Conscious Machines, and Animals: Broadening AI Ethics conference at Princeton where Constance and other attendees first learned about LLM's having bias against certain species and paying attention to the neglected topic of alignment of AGI towards nonhuman interests. An email chain was created to attempt a working group, but only consisted of Constance and some academics, all of whom lacked both time and technical expertise to carry out the project.  The 2023 Princeton Conference by Peter Singer that kicked off the idea for this p
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na