9fans - fans of the OS Plan 9 from Bell Labs
 help / color / mirror / Atom feed
* [9fans] fs performance
@ 2011-01-09 17:06 erik quanstrom
  2011-01-09 17:29 ` ron minnich
  0 siblings, 1 reply; 27+ messages in thread
From: erik quanstrom @ 2011-01-09 17:06 UTC (permalink / raw)
  To: 9fans

the new auth server, which uses the fs as its root rather than
a stand-alone fs,  happens to be faster than our now-old
cpu server, so i did a quick build test with a kernel including
the massive-fw myricom driver.  suspecting that latency kills
even on 10gbe, i tried a second build with NPROC=24. a
table comparing ken fs, fossil+venti, and ramfs follows.
unfortunately, i was not able to use the same system for the
fossil+venti tests, but there's a ramfs test on the same system
to bring things into perspective due to the large differences
in processor generation, network, &c.  here's an example test:

	tyty; echo $NPROC
	4
	tyty; time mk>/dev/null && mk clean>/dev/null
	2.93u 1.30s 3.36r 	 mk
	tyty; NPROC=24 time mk >/dev/null && mk clean>/dev/null
	1.32u 0.22s 2.29r 	 mk

and here are the compiled results:

a	Intel(R) Xeon(R) CPU           X5550  @ 2.67GHz
	4 active cores (8 threads; 4 enabled);
	http://ark.intel.com/Product.aspx?id=35365
	intel 82598 10gbe nic; fs has myricom 10gbe nic; 54µs latency
b	Intel(R) Core(TM)2 Quad CPU    Q9400  @ 2.66GHz
	4 active cores (4 threads; 4 enabled);
	http://www.intel.com/p/en_US/products/server/processor/xeon5000/specifications
	intel 82563-style gbe nic; 70µs latency

mach	fs	nproc	time
a	ken	4	2.93u 1.30s 3.36r 	 mk
		24	1.32u 0.22s 2.29r 	 mk
	ramfs	4	3.10u 1.67s 3.01r 	 mk
		24	2.98u 1.23s 2.42r 	 mk
b	venti	4	2.65u 3.44s 21.46r 	 mk
		24	2.98u 3.56s 21.58r 	 mk
	ramfs	4	3.55u 2.22s 9.08r 	 mk
		24	3.50u 2.67s 9.41r 	 mk

it's interesting that neither venti nor ramfs get any faster
on machine b with NPROCS set to 24, but both get
faster on machine a and the fastest time of all is not
ramfs, but ken's fs with NPROC=24.  so i suppose the
64-bit question is, is that because moving data in and
out of user space is slower than 10gbe, or because ramfs
is single threaded and slow?

in any event, it's clear that if the fs is good, latency
can kill even on 10gbe lan.  it would naively seem to me that
using the Tstream model would be too expensive, requiring
thousands of new streams, and require modifying at
least 8c, 8l, mk, rc, awk (what am i forgetting?).  but
it would be worth a test.

- erik



^ permalink raw reply	[flat|nested] 27+ messages in thread
* Re: [9fans] fs performance
@ 2011-01-10  3:57 erik quanstrom
  0 siblings, 0 replies; 27+ messages in thread
From: erik quanstrom @ 2011-01-10  3:57 UTC (permalink / raw)
  To: 9fans

> Peak Local file access bandwidth is typically 50 to 100 MBPs
> x number of disks; over the localnet it is about 80MBps. On
> my internet connection I barely get 1MBps download (& 0.2MBps
> upload) speeds.

interesting observation: when i first set up the diskless fileserver
at coraid, we had a mirror of the worm in another building across
an awful wireless connection.  we had no more than 1mbit.
at first i was a little worried about this, but then i realized that
128k/s * 86400s/day is 10.5gb/day.

btw, with aoe, you should saturate the network—125 mb/s/interface
for gbe so a typical el-cheepo computer these days can do 250mb/s
over aoe without breaking a sweat.  of course you'll get 10x that with
10gbe.

i agree with charles, network attached, or even internet-attached
storage seems like the way to go.  for internet-attached storage,
the amazing imbalances of very slow last-mile networks/very cheep
mass storage and the power of slow networks over time lead me
to think there are some very interesting engineering tradeoffs to
be made.

- erik



^ permalink raw reply	[flat|nested] 27+ messages in thread

end of thread, other threads:[~2011-01-11 11:33 UTC | newest]

Thread overview: 27+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2011-01-09 17:06 [9fans] fs performance erik quanstrom
2011-01-09 17:29 ` ron minnich
2011-01-09 17:51   ` John Floren
2011-01-10 18:07     ` Francisco J Ballesteros
2011-01-10 18:48       ` hiro
2011-01-10 19:06         ` erik quanstrom
2011-01-10 19:53         ` John Floren
2011-01-11 11:33           ` hiro
2011-01-09 18:31   ` erik quanstrom
2011-01-09 19:54   ` Bakul Shah
2011-01-09 20:25     ` ron minnich
2011-01-09 20:47       ` erik quanstrom
2011-01-09 21:04         ` ron minnich
2011-01-09 21:17       ` Bakul Shah
2011-01-09 21:59         ` erik quanstrom
2011-01-09 22:58         ` Charles Forsyth
2011-01-09 22:55           ` ron minnich
2011-01-09 23:50             ` Charles Forsyth
2011-01-10  3:26           ` Bakul Shah
2011-01-09 21:14     ` erik quanstrom
2011-01-09 21:38       ` Bakul Shah
2011-01-09 21:56         ` ron minnich
2011-01-09 22:02           ` erik quanstrom
2011-01-10 14:45           ` David Leimbach
2011-01-10 15:06             ` Charles Forsyth
2011-01-09 22:00         ` erik quanstrom
2011-01-10  3:57 erik quanstrom

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).