Create a gist now

Instantly share code, notes, and snippets.

What would you like to do?
Latency Numbers Every Programmer Should Know
Latency Comparison Numbers
L1 cache reference 0.5 ns
Branch mispredict 5 ns
L2 cache reference 7 ns 14x L1 cache
Mutex lock/unlock 25 ns
Main memory reference 100 ns 20x L2 cache, 200x L1 cache
Compress 1K bytes with Zippy 3,000 ns 3 us
Send 1K bytes over 1 Gbps network 10,000 ns 10 us
Read 4K randomly from SSD* 150,000 ns 150 us ~1GB/sec SSD
Read 1 MB sequentially from memory 250,000 ns 250 us
Round trip within same datacenter 500,000 ns 500 us
Read 1 MB sequentially from SSD* 1,000,000 ns 1,000 us 1 ms ~1GB/sec SSD, 4X memory
Disk seek 10,000,000 ns 10,000 us 10 ms 20x datacenter roundtrip
Read 1 MB sequentially from disk 20,000,000 ns 20,000 us 20 ms 80x memory, 20X SSD
Send packet CA->Netherlands->CA 150,000,000 ns 150,000 us 150 ms
1 ns = 10^-9 seconds
1 us = 10^-6 seconds = 1,000 ns
1 ms = 10^-3 seconds = 1,000 us = 1,000,000 ns
By Jeff Dean:
Originally by Peter Norvig:
Some updates from:
'Humanized' comparison:
Visual comparison chart:
Animated presentation:

need a solar system type visualization for this, so we can really appreciate the change of scale.

jboner commented May 31, 2012

I agree, would be fun to see. :-)

pmanvi commented May 31, 2012

useful information & thanks

dakull commented May 31, 2012

Looks nice kudos !
One comment about the Branch mispredict: if the cpu architecture is based on P4 or Bulldozer that would result in 20-30+ cycles on a mispredict that would translate to a much bigger number (and they do mispredict) :)

For SSD's would be something like:
Disk seek: 100 000 ns

Latency numbers between large cities:

@preinheimer Asia & Australasia have it bad.

Eronarn commented May 31, 2012

"Latency numbers every programmer should know" - yet naturally, it has no information about humans!

maybe you want to incorporate some of this:

Curious to see numbers for SSD read time

I think the reference you want to cite is here:

This remind me of this Grace Hopper's video about Nanoseconds. Really worthy.

mikea commented May 31, 2012

I find comparisons much more useful than raw numbers:

I'm surprised that mechanical disk reads are only 80x the speed of main memory reads.

dakull commented May 31, 2012

my version : includes SSD number, would love some more

Does L1 and L2 cache latency depends on processor type? and what about L3 cache.

dakull commented May 31, 2012

Ofc it does ... those are averages I think.

Would be nice to right-align the numbers so people can more easily compare orders of magnitude.

jboner commented May 31, 2012

Good idea. Fixed.

jhclark commented May 31, 2012

And expanded even a bit more: (SSD numbers, relative comparisons, more links)

TLB misses would be nice to list too, so people see the value of large pages...

Context switches (for various OSes), ...

Also, regarding packet sends, that must be latency from send initiation to send completion -- I assume.

If you're going to list mutex lock/unlock, how about memory barriers?

Thanks! This is quite useful, particularly for flogging at others.

lry commented Jun 1, 2012

Quick pie chart of data with scales in time (1 sec -> 9.5 years) for fun.

Spreadsheet with chart

"Read 1 MB sequentially from disk - 20,000,000 ns". Is this with or without disk seek time?

pgroth commented Jun 1, 2012

I made a fusion table for this at:

Maybe be helpful for graphing, etc. Thanks for putting this together

jboner commented Jun 1, 2012

Cool. Thanks.
Thanks everyone for all the great improvements.

ayshen commented Jun 1, 2012

Here is a chart version. It's a bit hard to read, but I hope it conveys the perspective.

It would also be very interesting to add memory allocation timings to that : )

How long does it take before this shows up in XKCD?

You guys are talking about is the powers of ten

If it does show up on xkcd it will be next to a gigantic "How much time it takes for a human to react to any results", hopefully with the intent to show people that any USE of this knowledge should be tempered with an understanding of what it will be used for--possibly showing how getting a bit from the cache is pretty much identical to getting a bit from china when it comes to a single fetch of information to show a human being.

@BillKress yes, this is specifically for Programmers, to make sure they have an understanding about the bottlenecks involved in programming. If you know these numbers, you know that you need to cut down on disk access before cutting down on in-memory shuffling.
If you don't properly follow these numbers and what they stand for, you will make programs that don't scale well. That is why they are important on their own and (in this context) should not be dwarfed by human reaction times.

@BillKress If we were only concerned with showing information to a single human being at a time we could just as well shut down our development machines and go out into the sun and play. This is about scalability.

klochner commented Jun 5, 2012

this is getting out of hand, how do i unsubscribe from this gist?

gemclass commented Jun 7, 2012

Saw this via @smashingmag . While you guys debate the fit for purpose, here is another visualization of your quick reference latency data with Prezi

Does anybody know how to stop receiving notifications from a gist's activity?

Here's a tool to visualize these numbers over time:

I just created flash cards for this: They can be downloaded using the Anki application:

I'm also missing something like "Send 1MB bytes over 1 Gbps network (within datacenter over TCP)". Or does that vary so much that it would be impossible to specify?

kofemann commented Feb 9, 2013

If L1 access is a second, then:

L1 cache reference : 0:00:01
Branch mispredict : 0:00:10
L2 cache reference : 0:00:14
Mutex lock/unlock : 0:00:50
Main memory reference : 0:03:20
Compress 1K bytes with Zippy : 1:40:00
Send 1K bytes over 1 Gbps network : 5:33:20
Read 4K randomly from SSD : 3 days, 11:20:00
Read 1 MB sequentially from memory : 5 days, 18:53:20
Round trip within same datacenter : 11 days, 13:46:40
Read 1 MB sequentially from SSD : 23 days, 3:33:20
Disk seek : 231 days, 11:33:20
Read 1 MB sequentially from disk : 462 days, 23:06:40
Send packet CA->Netherlands->CA : 3472 days, 5:20:00

kofemann commented Feb 9, 2013

You can add LTO4 tape seek/access time, ~ 55 sec, or ns

I'm missing things like sending 1K via Unix pipe/ socket / tcp to another process.
Has anybody numbers about that?

@metakeule its easily measurable.

mnem commented Jan 9, 2014

Related page from "Systems Performance" with similar second scaling mentioned by @kofemann:

izard commented May 29, 2014

L1D hit on a modern Intel CPU (Nehalem+) is at least 4 cycles. For a typical server/desktop at 2.5Ghz it is at least 1.6ns.
Fastest L2 hit latency is 11 cycles(Sandy Bridge+) which is 2.75x not 14x.
May be the numbers by Norwig were true at some time, but at least caches latency numbers are pretty constant since Nehalem which was 6 years ago.

richa03 commented Aug 21, 2014

Please note that Peter Norvig first published this expanded version (at that location - ~JUL2010 (see wayback machine). Also, note that it was "Approximate timing for various operations on a typical PC".

pdjonov commented Oct 3, 2014

One light-nanosecond is roughly a foot, which is considerably less than the distance to my monitor right now. It's kind of surprising to realize just how much a CPU can get done in the time it takes light to traverse the average viewing distance...

junhe commented Jan 16, 2015

@jboner, I would like to cite some numbers in a formal publication. Who is the author? Jeff Dean? Which url should I cite? Thanks.

I'd like to see the number for "Append 1 MB to file on disk".

The "Send 1K bytes over 1 Gbps network" doesn't feel right, if you were comparing the 1MB sequential read of memory, SSD, Disk, the Gbps network for 1MB would be faster than disk (x1024), that doesn't feel right.

leotm commented May 2, 2015

A great solar system type visualisation:

ali commented Sep 14, 2015

I turned this into a set of flashcards on Quizlet:

Can you update the the Notes section with the following
1 ns = 10^-9 seconds
1 ms = 10^-3 seconds


jboner commented Dec 13, 2015

@misgeatgit Updated

Zippy is nowadays called snappy. Might be worth updating. Tx for the gist.

Several of the recent comments are spam. The links lead to sites in India which have absolutely nothing to do with latency.

Are there any numbers about latency between NUMA nodes?

vitaut commented Jan 31, 2016

Sequential SSD speed is actually more like 500 MB/s, not 1000 MB/s for SATA drives (,3269.html).

You really should cite the folks at Berkeley. Their site is interactive, has been up for 20 years, and it is where you "sourced" your visualization.

Question~ do these numbers not vary from one set of hardware to the next? How can these be accurate for all different types of RAM, CPU, motherboard, hard drive, etc?

(I am primarily a front-end JS dev, I know little-to-nothing about this side of programming, where one must consider numbers involving RAM and CPU. Forgive me if I'm missing something obvious.)

The link to the animated presentation is broken, here's the correct one:

keenkit commented Aug 15, 2016

Love this one.

profuel commented Oct 5, 2016

Mentioned gist : is private or was removed.
can someone restore it?

trans commented Oct 9, 2016

It would be nice to be able to compare this to computation times -- How long to do an add, xor, multiply, or branch operation?

mpron commented Oct 12, 2016 edited

Last year, I came up with this concept for an infographic illustrating these latency numbers with time analogies (if 1 CPU cycle = 1 second). Here was the result:

pawel-dubiel commented Jan 29, 2017 edited

Most of these number were valid in 2000-2001, right now some of these numbers are wrong by an order of magnitude. ( especially reading from main memory, as DRAM bandwidth doubles every 3 years )

µs, not us

GLMeece commented Jan 31, 2017 edited

I realize this was published some time ago, but the following URLs are no longer reachable/valid:

However, the second URL should now be:

Oh, and @mpron - nice!

Thank you @jboner

GLMeece commented Jan 31, 2017

Note: I created my own "fork" of this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment