Skip to content(if available)orjump to list(if available)

You Don't Have Time Not to Test

You Don't Have Time Not to Test

69 comments

·April 2, 2025

gcp123

This struck a nerve.

I've been on both sides of this war: the test evangelist fighting for coverage and the pragmatist shipping to beat a deadline. After 20+ years in software, the truth is painfully obvious: testing is the greatest productivity hack that everyone keeps "postponing until next sprint."

The author gets the psychology exactly right. We overestimate the initial cost and drastically undervalue the compound returns. What they call "Time Technical Debt" is the perfect description for that sinking feeling when you're working on a mature codebase with spotty test coverage.

The most insightful point is how testing fundamentally changes your design for the better. When you have to make something testable, you're forced to:

- Think about clear interfaces

- Handle edge cases explicitly

- Create clean separation of concerns

- Build proper startup/shutdown sequences

These aren't "testing best practices," they're just good engineering. Testing is simply the pressure that forces you to do it right.

My experience: if your system is hard to test, it's probably hard to reason about, hard to maintain, and hard to extend. The difficulty in testing is a symptom, not the disease.

At my last company, we built a graph of when outages occurred versus test coverage by service. The correlation was so obvious it became our most effective tool for convincing management to allocate time for testing.

magicalhippo

While I agree testing in general is a must, I'm still not sold on unit testing as a general tool.

For stuff like core libraries or say compilers, sure, unit tests are great. But for the levels above I'm leaning towards integration tests first, and then possibly add unit tests if needed.

After all, you can't not have integration tests. No matter how perfect your Lego bricks are, you can still assemble them together the wrong way around.

mewpmewp2

I tend to think that the most valuable tests are on the exact level of when something gets used by multiple consumers.

So shared functions get unit tested, but if there is a non shared function that gets triggered only from few layers up via user click, it is better to simulate that click and assert accordingly. However the exceptions are when the input can have tons of different permutations, when unit test might just become more optimal then.

jasonjmcghee

Yes exactly. If you're writing a library, unit tests are likely a great investment. Your "customers" are using the things you are testing.

If you're writing an API, your customers are calling the endpoints. Very important they behave properly on happy path and fail appropriately when they should (permissions / security especially)

(I also ensure tests in the service/dao layer to protect developers from footguns and/or more "unexpected" behavior, but I'd argue the API layer is generally more important)

If you're writing a react app, playwright is very important (and using it properly / acting like a real user) as your customers are interacting with it via a browser.

Though the larger the codebase/ when you start having reusable components, the better tested they are, the less will go wrong in the future as an unfamiliar developer will inevitably hold it wrong.

roland35

I think in general tests should be where errors and mistakes are more likely to occur. Different code bases could be different! Hard core math libraries are different than a web app with various integrations.

umvi

Agree - unit tests are best for pure functions and the like. If you are having to do a ton of mocking and injection in order to unit test something, it's probably a sign black box testing might be higher value

jasonpeacock

Integration test are slow/expensive to run compared to unit tests and reduce your iteration speed.

Unit tests let you change code fearlessly with instant feedback.

Integration tests require basically deploying your app, and when something fails you have to debug the root cause.

If you’re doing a lot of mocking then your design is not good. And only public interfaces should have testing.

maccard

On every large codebase I’ve worked on , updating a low level function has required more work updating the tests than updating the application using it.

Unit tests have a place, but IME are overused as a crutch to avoid writing useful bigger tests which require knowing what your app does rather than just your function.

> Integration test are slow/expensive to run compared to unit tests and reduce your iteration speed.

My unit tests might take under a second to run but they’re not the ones (IME) that fail when you’re writing code. Meanwhile, integration tests _will_ find regressions in your imperfect code when you change it. I currently use c# with containers and there’s a startup time but my integration tests still run quickly enough that I can run them 10s or hundreds of times a day very comfortably.

> If you’re doing a lot of mocking then your design is not good.

This is the “you’re holding it wrong” argument. I’ve never seen a heavily unit tested codebase that didn’t either suffer from massive mocking or wasn’t decomposed into such small blocks that they were illogically small and the mental map of the project was like drawing with sand - completely untenable.

> And only public interfaces should have testing.

This is the smoking gun in your comment - I actually disagree and think you should infer this. Services (or whatever you call them) should be tested, and low level functionality should be tested but the stuff in the middle is where you spend 80% of your time and get 10% of the benefit

magicalhippo

> Unit tests let you change code fearlessly with instant feedback.

Sure they can add confidence in making changes. But I've seen they can also give you false confidence.

As I said, I can still assemble your perfect Lego bricks together wrong. And you can still change the public behavior while keeping your unit tests passing fine. I've seen both happen in practice.

That's why I think integration tests (or whatever you want to call them) give you more bang for your buck. They give you even greater confidence that your stuff works, and generally you can cover a lot more with far fewer tests so improves ROI.

The tradeoff is that it can take a bit longer to run.

> If you’re doing a lot of mocking then your design is not good.

If my app needs to talk to a database, store things in object store, send some messages in a message queue and so on, I can't not mock those things away if I'm to write unit tests.

MoreQARespect

>The most insightful point is how testing fundamentally changes your design for the better. When you have to make something testable.

When people say this type of thing I consider it to be kind of a code smell that they're testing at too low a level and tightly coupling their tests to their implementation.

It is true that the pain of tightly coupling your code to your tests to your implementation can drive you to unwind some of that tight coupling but that still leaves your tests and your code tightly coupled.

I find the best bang for my buck are tests that are run at a high enough level to make it possible to refactor a lot and safely with a minimum of test changes. Technical debt that is covered by tests doesnt compound at nearly the same rate.

paulryanrogers

IME end-to-end tests in a browser really help with services that have a lot of parts to integrate, but damn they are hard to make reliable.

One challenge is animation and timing races, supposedly Playwright can address many of those. Another is some infrastructure like GitHub Actions can be randomly resource starved, such as causing the Chrome Driver to become unresponsive. Automated retrying is one workaround, at the cost of possibly papering over rare race and timing issues.

Of course unit tests are nice and fast and narrow. But refactors could render a large portion obsolete, and they won't prove things work together as a whole.

soneca

Do you think early startup product might be an exception?

Since whole features can be quickly ditched frequentl. Sometimes even complete product pivot.

edoceo

No excuses. I've done this like 12 times as startup CTO.

The habit is important. If you don't start, after three pivots you'll have a huge mountain of tests for a system nobody understands. Plus all the wasted time manually "testing".

Tests are so critical for the success of the business. It's fiscally irresponsible to skip.

mewpmewp2

What about side projects that you are working alone on?

wavemode

No, I don't think there is any exception. If you intend to maintain a piece of software for any length of time (i.e. it's not just a throwaway demo), you should write tests for it.

Over time you realize that testing truly does not slow down development as much as many people think it does. Maybe devs who just aren't used to testing find it difficult, but after a while it becomes second nature.

The best thing an early startup CTO can do is enforce testing across the board, so people don't just test when they feel like it.

leptons

>Over time you realize that testing truly does not slow down development as much as many people think it does.

Not my experience. We just built a new codebase, rewriting an older project with typescript and all the modern libraries and conveniences. We spent about 2x more time writing the tests than we did any of the API code.

Tests can be so fiddly and not exactly straight-forward. It takes a lot of time, but that isn't a reason not to do it. But don't suggest it's going to take less time, even in the long run, because it isn't - you essentially have to maintain 2 codebases now, one for the actual code, and one for the tests. Both are points of failure and both can be a time-sink.

TheCoelacanth

Only if your runway is measured in days rather than weeks or months.

The payback for good testing is very fast, especially once you have set it up for the first feature.

shanemhansen

For me personally tests have a positive ROI within hours.

Even if I was doing a one day hackathon I'd probably have some sort of test feedback loop.

I've dealt with P1 bugs that cost the company 100k/minute and still took the time to write a test for the fix because you really don't have time to get the fix wrong and not find out until it is deployed.

wombatpm

Whenever I’m told there isn’t time, I remind the PM :Yes, but if it’s wrong or doesn’t work we will have time to do it over - it will just cost more later.

leptons

I an early start-up is the exception, but my boss didn't. We were still in "stealth mode" and the CTO wanted 100% test coverage on our nodejs based social website, from the very start. 6 months in and we didn't have all that much built, because they couldn't really decide what they wanted us to build. So we built the most well-tested email sign-up form that ever existed, and a bunch of other user-account related stuff too, but then the company completely pivoted at around 6 months and I was now somehow doing PHP programming (which I hate) hacking the code of some ad server and bolting it on to a mobile app (not what we set out to build), and at that point the requirement for tests had been forgotten, because the company was desperate to find any viable path forward. It dissolved about 3 months after that, and now those tests seem pretty pointless.

TheCoelacanth

100% coverage is a vanity metric and a waste of time.

Focus on testing the use cases that are important to users, not on covering every single line.

Fire-Dragon-DoL

I think testing as a cost and manual QA has another cost. Depending on the cost over time (maintenance or manual QA) and the upfront cost, as well as the risk of refactors, which needs tests, there are moments where one option is better than the other. It changes, so it is worth considering.

hobs

I dont think they replace each other, they are complementary - they do overlap though, which is why people think you can trade them off - a test is not creative, a human is not an automaton.

Fire-Dragon-DoL

We have been using humans for automations for a long time, factories have been a thing for quite a while now.

They have an intersection, the fact is that you might need only the benefits in the intersection, not all the benefits. Given that, it could make sense.

I have way more fun writing tests, but automating certain things have a steep maintenence cost, a checklist for humans is a better idea in that case

klysm

- Think about clear interfaces

- Handle edge cases explicitly

- Create clean separation of concerns

- Build proper startup/shutdown sequences

If you do all of these things to start with though, then what's the value proposition?

t-writescode

Refactor simplicity, regression reduction, reduced time to “next launch” because manual validation period is shorter.

Increased customer trust because fewer regressions get missed.

bdangubic

how do you know that you did:

- thought about clear interfaces?

- created clean separation of concerns?

- built proper startup and shutdown sequences?

:)

simonw

If the way you write automated tests is slowing you down rather than speeding you up, you need to rethink the way you are writing tests.

In my opinion tests are the ultimate productivity hack for building software. In exchange for a relatively small up-front cost (if you know what you are doing) they let you move so much faster! You can make changes with confidence that each change won't break your existing functionality, without having to spend hours on manual testing for every change you want to ship.

This goes for smaller personal projects too. I gave a talk about this a few years ago, originally titled "Massively increase your productivity on personal projects with comprehensive documentation and automated tests" but then renamed (after I'd pitched the talk but before I presented it) to "Coping strategies for the serial project hoarder". Video, slides and notes here: https://simonwillison.net/2022/Nov/26/productivity/

mewpmewp2

I totally understand writing tests on professional projects and on projects that I am working with others, but for my side projects, I could never. I don't feel like I have lost anything by not writing tests in my side projects.

My issue with tests in my side projects is that it kills my flow. I need to actively iterate on the code that matters and feel out the results. At what point should I be writing tests?

I can only think of a point where I am certain this is where I am going to stay at and then I write some sanity regression tests.

simonw

It is deeply unintuitive that writing tests for personal / side projects would be worthwhile. I found that my productivity on those projects increased dramatically when I adopted the same testing regime I had previously used in my day-job with an engineering team distributed across multiple continents. That's why I wrote that talk!

"At what point should I be writing tests?"

Once the code does the right thing, and you want it to keep doing the right thing. Don't try to write the tests first, that's a productivity anti-pattern.

pydry

>Don't try to write the tests first, that's a productivity anti-pattern.

Uh, this isnt great advice.

If you're spiking then sure, dont write a test at all.

If you're writing production code you should have a clear picture of what it should do before you start. Your test should validate that so it makes sense to write that next. It serves as a natural transition from your requirement to youe code.

Ive written a lot of code without tests at all for good reasons and a lot of tests before code but ive never witnessed a good reason to write a test after writing the code. At best it is almost as beneficial.

kragen

Sometimes that happens, because the difficulty is not in getting it to do what you want, but figuring out what you want. That's the main theme of my top-level comment; I'm interested to hear what you think. https://news.ycombinator.com/item?id=43597869

deterministic

Completely agree. Having great automatic tests makes you move so much faster and drop fixes into production with justified confidence.

angarg12

So I joined this company as a lead of a team of 4 developers. I found a medium size Java codebase with not a single test. Instead of unit tests, we had a list of scenarios, and at the end of each sprint a QA guy would manually go through each scenario and verify the software.

One of my first moves was to ask the developers to write tests for their code. I got terrible pushback, specifically from one of them that said the dreaded "we don't have time to write tests".

This went back and forth a few times until it got escalated to the lead of the whole project, who sided with the developer: writing tests takes too much time and we don't have enough.

Two days. It took me two days of my own time to go through a completely unfamiliar codebase and get a reasonable code coverage with unit tests.

The benefits were immense and immediate. For starters, we caught a bunch of regressions before a new release that might have taken days of manual testing by QA. We were also able to ship faster and with more confidence.

As OP say, if you don't have time to test, you certainly don't have time not to test.

wdpk

During my several decades of translating businesses logic into code I never found a test that does not pay back the time spent to write it many times over, they are truly the super-power of the trade mainly because the can be completely automatized. Because of this automatic nature from a business perspective their marginal cost becomes almost zero (pretty much regardless of the initial cost and maintenance cost), the reason for this is that the value function of the business logic embedded in code, the reason for the code to even exist (i.e. the thing that gives a value to the piece of software), needs to account in the long run for the risk components, risks generated by defects or poorly aligned specifications, security risk, reputation risk, obsolescence, performance, etc. The way to manage risk at the business level is by somehow hedging and this hedge is precisely testing, again and again, hopefully automatically. Testing is then not a cost, it's precisely the instrument that allows to stop paying over and over for those risks and therefore in the end capture more value and reduce the variance of the outcomes.

People also mentioned side projects and testing. What I found personally is that tests are the best way to also document/encode ideas for later so that when you pause something for months because... life, getting back to a project becomes easier. I found very useful for instance to add some tests that would just fail for some feature easy to implement just to have an easy entry point back into the project.

ashishb

Tests can be good. Tests can be bad.

  - If you are writing a heavy algorithmic code, then write unit tests.
  - If you are writing glue code joining services, then write integration tests.
  - Integrating with third-party services means you prefer circuit-breakers and production monitoring over tests.
The tests should ideally be written along the axis of minimum change. Or the test code churn alone would slow down the product.

jt2190

> “We don’t have time to [write automated] test[s],” stated the principal engineer… This was in response to my candid observation that test coverage was low, existing tests didn’t always run locally for mysterious reasons, integration tests were non-existent, and pull requests could merge without passing the few tests that did exist.

Given the state of the existing automated tests, perhaps the Principal Engineer is correct… Why would the new automated tests be any better?

default-kramer

Yeah, automated tests are great when they are good. But when they are bad, they can be worse than no tests at all.

Multiple times in my career I've taken ownership of a codebase written by someone else. Some of the most pleasant ones had very little test coverage, but the code quality was high enough to make the lack of tests not really matter. On the other hand, some code is so bad covering it with tests adds nothing of value; it's just one more thing I have to analyze and decide what to do with.

I agree with most of the author's points in general, but I think that most programmers who are capable of writing good tests are also capable of recognizing situations when it's not worth the trouble.

Scene_Cast2

I'd love to hear how people write tests for ML. When I'm doing a greenfield project with a new model, a lot of issues are very statistical, e.g. incorrect downsampling - the model will run and train, just less optimally than normal.

I can't put optimality bounds because I don't know how well the model _should_ train, and when it doesn't train, that's not necessarily because of an incorrect implementation. And, actually training a model for a test is quite resource and time heavy.

t-writescode

You can test:

  * the tools that operate the model.
  * to make sure the fitness function calculates fitness correctly,
  * simulation runs right.

  * the storing and recovering of the model operate correctly - that is, data is saved and then recovered correctly and consistently.
  * that the engine that runs the training operates correctly
  * that shutdown and startup work
  * the a crash can be recovered from usefully
And I’m sure more

roland35

Tests are a superpower with hardware development. It can be difficult to find the best boundaries to mock, but once you do it's amazing how much easier it is to develop! I find that tests are a great way to get the software into states that are hard to recreate in real life, such as faults as well

kragen

Sometimes you're writing code because you want to see what happens. Richard P. Gabriel wrote a wonderful, moving essay about this recently, "I Throw Itching Powder at Tulips": https://www.dreamsongs.com/Files/Tulips.pdf

> Software engineering is not what I do when I program. I am programming. I write software as part of doing science. I use software as a machine or instrument to explore how the mind / brain might work. Not when the mind’s thinking—that’s old AI. I mean when people are creating.

I find that it's pretty common when I'm, say, tweaking gameplay in a game, that I have no trouble making the code behave the way I expect, but impossible to predict ahead of time how I will feel about different possible behaviors. I have to try them on to see how I feel in them. Adding tests to that feedback loop doesn't help; it just slows it down, dramatically slowing down the process of learning and exploration.

On the other hand, when I am writing code to meet someone else's needs rather than as a form of self-expression or exploration, I usually want the code to behave in a specific way, at least in certain situations, and to keep behaving that way as I make other changes to it in the future, unless I change my mind. In those cases I think it's very rare for automated tests to slow me down instead of speeding me up.

In languages with expressive static type systems like OCaml, the crossover happens a little later, because the static type system already constrains the program's behavior a lot. But automated tests still become very valuable very soon when I want a specific behavior.

I've also, unfortunately, worked on projects where the test suite we had was worse than nothing. Invariably it wasn't because we were throwing itching powder at tulips so that the optimal test suite would be empty; it was because we'd done a bad job writing the tests.

simonw

This is why I'm not at all a fan of pure test-driven development where you try to write the tests before the implementation. I think that's often a big time sink and can put people off writing tests entirely.

Instead, I like tests-in-the-same-commit development. I write the implementation, tweak it until it works, then add a test that proves it does what it's supposed to before I commit the code. I wrote about that here: https://simonwillison.net/2022/Oct/29/the-perfect-commit/

I often use snapshot testing for this, which is very fast to implement. I have a short note about one way to do that here: https://til.simonwillison.net/pytest/syrupy - and a longer piece here: https://simonwillison.net/2020/Feb/11/cheating-at-unit-tests...

kragen

I have found test-first development valuable sometimes. It's a proper subset of the times that automated testing in general is worthwhile; there's fairly strong empirical support in the academic literature for automated testing improving productivity and quality, but test-first seems to be a wash overall.

Test-first development has some advantages in my experience:

- You know the test isn't vacuous because you saw it fail when you first ran it.

- You know you have thorough test coverage because you stopped adding code to the implementation when all the tests passed (at least temporarily).

- You know the interface you're implementing isn't a huge pain in the ass to use because you would have noticed that halfway through writing the test case and redesigned it.

- You have a shorter feedback loop between writing the implementation and seeing a test start passing, which can be a useful motivation hack and can also speed up your learning of the language and libraries you're using to build the implementation.

But there are also disadvantages:

- You can't use snapshot testing. I didn't know about syrupy, but doctest is also very amenable to snapshot testing.

- The rapid feedback loop can be a motivational trap, getting you lost in tactical details (as Ousterhout famously worried in his debate with Martin) at the expense of attention to more strategic concerns that are ultimately more valuable.

- Sometimes it's a bigger concern to define an interface that it's feasible to implement than one that's a pleasure to use. Test-first development backfires in that case, since first you wrote a test that's impossible to pass, then an article implementation that doesn't pass it, and then a second test that are tests the actually feasible implementation. A common case of this is where you left some necessary resource or information out of the interface.

With respect to your piece about snapshot testing, which is excellent, I actually disagree about using tests to verify security properties. I mean, you should definitely have automated security tests, yes, but you should never rely on them. Code reviews, logical arguments, and ideally formal proofs are much better. Attackers can be counted on to try things you didn't think about when you were writing the tests. Randomized testing like Hypothesis or AFL can help, but not enough.

But, at the other extreme, there are some programs for which automated testing itself is of no value.

simonw

I find test-first development useful for very specific things. If I'm building a Markdown parser for example, where the inputs and outputs are very clear, I may break out test-first development.

In place of test-first for test verification I'll often deliberately break my new test (either in the test or the implementation), run it once to make sure it fails, then fix it again.

m463

I remember asking folks for tests for their stuff, and they didn't want to do it because it would take too much time and effort. And I said "It doesn't have to be perfect, just do a simple sanity check".

It's possible people on the opposite ends of the perfectionism scale might view tests differently.

The perfectionists want 100% test coverage which takes a serious commitment of time and effort, while the fast-and-loose folks don't want to bother with tests at all.

I wonder if the best way to test is to have the fast-and-loose people make fast-and-loose tests, and then let the perfectionists iterate on the tests as time goes on.

simonw

Totally agree. One of my favourite style of tests is this:

  for path in ("/", "/news/", "/about/"):
      response = client.get(path)
      assert response.status_code == 200
Run that on every commit and if you accidentally ship something that throws an exception on one of those pages you'll know about it instantly.

3eb7988a1663

I have spotty test coverage, but always configure a high level smoke test on every end point. Prevents silly mistakes that might otherwise slip through and trivial to implement. Can supplement for the more complex interactions, but gives a big safety blanket that there has not been a regression which breaks everything.

PaulHoule

Depends on your toolset. For some kind of code, testing is the solution and not a problem. For React before version 17 between the test code never being able to know if your callbacks have settled and react-testing-library insisting that you aria-everything or screw up your code with test id’s, writing a test is hours of wasted time now and more hours wasted gradually over the next year because your tests already take 70 sec to run. React always knew what was queued to run and why it took them 17 releases to expose it to test runners is beyond me.

wewewedxfgdf

Context matters.

If you are a startup without customers/users then writing tests is a bad idea - you need to focus everything on getting working code and users. You may be throwing your code away because you still don't even know what your users really want.

On the other hand if you are writing a banking system then sure, testing is not negotiable.

"Should we be writing tests?" cannot be answered without also knowing what the software is, what the company does and the context in which the software is being written.

simonw

Built well, tests can accelerate your pace of development even before you have launched to real users.

You don't need 100% coverage, but having tests that help spot dumb mistakes you've made (like a 500 on your homepage) are cheap to implement and pay for themselves very quickly.

t-writescode

If you’re a startup handling customer data, how can you skip out on tests? How can you be confident that your code doesn’t expose customer data in an exciting way and in a that demolishes trust in your product without testing to make sure and adding continuous tests to make sure a mistake like that doesn’t make it live?