Skip to main content

Kitchen Cabinets

Imagine a scenario where you're helping a friend pack up their kitchen:

  • The kitchen contains 10 cabinets.
  • A cabinet takes one hour to pack.
  • However, each cabinet may have a further cabinet nested within it. And on ad infinitum.
  • There is a 1:5 chance of a cabinet containing another cabinet.

How long should you estimate for the job? (The answer is below)

This was suggested in a Hacker News comment discussing software estimation, and struck a chord with many readers. It's clear that we are no longer in the Fill-The-Bucket domain anymore; our original intuitions about how long things might take are not going to work here.

As a developer, this 'feels' more real to me than Fill-The-Bucket. Any task I take on has an outside chance of telescoping into something much worse. Here's a recent example:

  • I wanted to test out a CSS change to my website site. 1 hour?
  • But in order to avoid wrecking the live version, I would need to test this offline, with Jekyll. 2 hours?
  • However, this meant getting Jekyll working locally on my mac (again). 3 hours
  • However, Jekyll didn't install, because it used a later version of Ruby than the one I had. 4 hours?
  • The Ruby version Manager (RVM) I had was out-of-date, and needed updating. 5 hours?
  • And I couldn’t update it because brew has been disabled by an update to my OS-X installation. ??? hours

All-in-all, it just may not have been possible for me to test that CSS change the way I wanted to. Every task had a further, worse sub-task embedded within it.

Cabinets within cabinets.

Distribution

The above chart simulates the kitchen cabinet scenario. Have a play and see how different numbers of cabinets and probabilities work out. In particular, what happens when:

  • You have thirty cabinets in the original kitchen?
  • You have a single cabinet in the original kitchen, and say a .8 chance-of-nesting?

When the number of initial cabinets is high, we are closer to the Fill-The-bucket world, with it's normal distribution, and variance-around-a-mean.

But when the number of initial cabinets is low, the distribution is "long-tailed" and tends towards the Exponential Distribution, which works in a way similar to radioactive decay. We might best be able to talk about moving kitchens in terms of their half-lives.

That is, given a bunch of infinity-cabinets, we could say how long it would usually take for half of them to be completed. Then, it'll be the same again for the next half, and so on.

Whereas Fill-The-Bucket was defined with a mean and variance, the exponential distribution is modelled with a single parameter, lambda (λ), which is the rate of decay.

As you can see from playing with the above chart, with low values of lambda, our completion time is much more likely to take longer.

But what (if anything) does this lambda parameter represent? It could be:

  • Unplanned re-work. Often, version one is lacking in features, and makes incorrect assumptions, which can only be unpicked by throwing away large chunks of work and starting again.
  • Unforeseen Dependencies. Like the above CSS/Jekyll/Ruby example, doing the work uncovers extra work that needs to be done.

Risk In An Exponential World

Let's assume that the exponential distribution does model software development. Where does that leave us with respect to estimating risk?

With any estimate, there are risks in both under- and over- estimating:

  • Too Long: In estimating too much time, you might not be given the work or your business might miss the opportunity in the marketplace. A too cautious risk might doom a potentially successful project before it has even started.

  • Too Short: If you estimate too little time, you might miss important coordinating dates with your marketing team, or miss the Christmas window, or run out of "runway".

For the sake of our simulation's simplicity, let's assume that being a day long on the estimate is as bad as being a day short and that the cost is linear in both directions.

See how this plays out in the following set of charts.

Analysis

We have four charts above:

  • The top (red) chart is our exponential probability density. Our eventual completion date is randomly picked from somewhere in the red area. The grey overlay shows the rate at which projects are likely to be completed.
  • The second (green and blue) chart shows the costs of getting the estimate wrong. If you hit the date exactly, you minimize risk. If you overrun, you pick up the green costs. If you finish early, you pick up the blue costs. You might question why you have costs for finishing early, but you are being penalized here because your estimate was wrong. Maybe the product could have shipped sooner if you'd given a more accurate date?
  • The third (orange) chart shows our risk profile - it's the (green and blue) costs from the middle multiplied by the probabilities (in red) from the top.
  • The bottom (green and blue) chart is the area under the risk profiles from the third chart. That is, we're providing a simple way to compare the early risk with the late risk.

Balancing Risk

Keeping lambda at .25, at four days, the too-early risk is the same as the too-late risk. This might seem like a good choice for the project.

But, it's not the point of lowest risk.

For that, you'll have to go somewhere around three days. This is roughly the mean length of the project - with this lambda, half of all projects would be completed before this point, and half after. But the too-late risk is much greater than the too-early risk.

So an important question at this point might be: who bears this risk? In the models above, we make the simplifying assumption that too-early and too-late risks have the same cost, but in reality, missing deadlines can be damaging to your personal reputation.

Meta-Analysis

Knowing Lambda

Is lambda predictable on a project? It doesn't appear that there have been any efforts in our industry to decide such a thing, even though we know that re-work and uncovering new problems happens. We also know that if we ship some code, there will be bugs in it. How many? How much work will they take to fix?

When Does Risk Happen?

Too-early and too-late risks are both [Scarcity Risks](/risks/(/tags/Scarcity-Risk): they reflect the fact that time/budget/staff/opportunity are scarce resources which you can run out of.

But where is the risk accrued? If you give an estimate, you lock in a maximum too-early risk at that point. From then on, the clock is ticking: too-early risk decreases towards zero as the due-date approaches.

This is important: the point at which you present your estimate is the point of highest too-early risk. It's also the point at which this risk is either accepted or rejected.

Accepting an estimate

The diagram above is an example of this. A supplier is bidding for a contract with a client. The client has functionality they want build (or Feature Risk as we call it on Risk-First). The supplier needs money to keep their business going (Funding Risk on this diagram).

If the estimate is accepted, the supplier's Funding Risk is transferred to the client (the requester of the estimate). Conversely, the trade is that the client's Feature Risk is transferred to the supplier.

If the supplier is short on opportunities or funds, there is a tendency to under-estimate. That's because the Feature Risk is a problem for the supplier in the future, whereas their Funding Risk is a problem right now.

You can often see suppliers under-bid on projects because of this future discounting, which we discussed before in Evaluating Risk.

This analysis also suggests something else: the process of giving and accepting estimates transfers risk. This is a key point which we'll return to later.

Too-Late Risk

Conversely, too-late risk accrues only after the delivery date has passed. Like too-early risk, there is probably a maximal limit on this too, which occurs at the point the project is cancelled due to lack of funds!

The problem with projects in the Kitchen Cabinet domain is that elapsed time is no indication of remaining time. The exponential distribution is exactly the same shape at every point in time (we're dealing with half-lives, remember?).

This means that clients often keep projects running for far longer than they should, assuming success is just around the corner. This is an example of the Sunk Cost Fallacy.

Being On-Time

There is an alternative to too-early or too-late risk. You can always choose to be on time. This is definitely a choice: Just like a student can always hand something in on assignment day (even if it's just a title scrawled on a piece of paper), you can always hand in whatever work you have.

Then, instead of worrying about [Scarcity Risks](/risks/(/tags/Scarcity-Risk), you are letting Feature Risk vary to take up the slack.

So far, we've seen two kinds of estimate: Fill-The-Bucket and Kitchen-Cabinet. Now, it's time to review a third - estimating Journey Style, and looking at how we can minimise Feature Risk within an available budget.