I just got a new shiny desktop at home, a Quad Core i7 920, running 64-bit Linux 2.6.32 (stock Fedora Core 12). Nothing fancy, it's a $250 CPU, but I did some artificial tests of Traffic Server on it just to see where we are on a modern machine and kernel. The tests are done between two Linux boxes, both on GigE network, and with two Linksys switches in between the two boxes. So, this is definitely not a production quality network in any way. I ran with keep-alive enabled, doing 100 requests per connection, each request fetches a 500 byte body out of cache (i.e. 100% cache hit ratio). I know, this is a completely unrealistic test, but it is still interesting to see what we can do. Here's the best run out of three:
2270442 fetches on 23464 conns, 2000 max parallell, 1.13522e+09 bytes in 30 seconds 500 mean bytes/fetch 75681.40 fetches/sec, 3.78406e+07 bytes/sec msecs/connect: 0.3878585 mean, 11.1295 max, 0.089 min msecs/first-response: 16.3695 mean, 459.379 max, 0.1295 min
So, over 75,000 requests per second, at 16ms latency, and I think some of that latency can be attributed to the two switches (wish I had a better setup). An interesting side note, the Traffic Server process can actually use about 470% CPU on this Quad Core box, so an extra 70% of a CPU is "gained" here from Hyper Threading. How does this compare to my old desktop? Well, the old system was a Core2, and in the same test, it was able to pull off around 35k QPS.
Anyways, these results aren't too shabby, and we've just started!
Caveat: This is using the "trunk" version of Traffic Server, the first release that is soon coming out won't be quite this fast.
Update: I updated the numbers with the results after forcing my CPU to run at highest CPU frequency at all time. No overclocking though, this is a standard i7 920 setup.