From: Erick Engelke on

Hi,

My team is looking at upgrading our FreeBSD fileserver, we've been
extrememly happy with FreeBSD, the fileserver serves:

umpteen NFS clients (all FreeBSD)
2,000 CIFS clients with Samba
9,000 homedirectories on about 4 TB of RAID disks on InforTrend RAID
chassies, 2GB battery backed up RAM on RAID chassis
with user quotas

Our present box is dated. It's FreeBSD 6.2, dual processor, 4GB RAM
running 32 bit OS. We will find some money to upgrade it.

We've looked at (and talked to) NetApp and Sun about replacement products,
but we're also evaluating an inhouse FreeBSD solution to support:

all 9,000 homedirectories on 16 TB RAID disks growable to 92 TB

ZFS looks nice, except that to support 9,000 user accounts, we believe we
would have to have 9,000 mount points. That's how the Sun system would
work.

The Sun and Netapp solutions have about 64 MB servers and a lot of SATA
drives. It looks like we could build a FreeBSD systems but with SAS for a
similar price. The way the users hammer our hard disks, I'm really not
thinking slow spinning, slower seeking SATA drives are a good idea, even
though I realize ZFS and oodles more RAM to cache would help performance.

Sun's Storage 7000 series family uses an 18 GB solid state hard disk which
buffers the ZFS writes and is supposed to really speed up the system. Can
I do the same with FreeBSD.

In the end, we may purchase one of these brand name systems, but I want to
give FreeBSD a fighting chance because it has been rock solid for all our
server systems. We started on FreeBSD 2.0, so we've seen a lot of
progress.

Erick

Erick Engelke erick(a)uwaterloo.ca
Director PHY-3013
Engineering Computing (519) 885-1211 x35893
University of Waterloo http://www.eng.uwaterloo.ca/~erick

From: Michel Talon on
Erick Engelke <erick(a)engmail.uwaterloo.ca> wrote:
>
> Sun's Storage 7000 series family uses an 18 GB solid state hard disk which
> buffers the ZFS writes and is supposed to really speed up the system. Can
> I do the same with FreeBSD.

I think you can, and this is the best option to speed up ZFS. You should
ask on a FreeBSD mailing list where developers hang up, such as
freebsd-stable. The real question is: what is the reliablity of ZFS for
such high intensity setup as you envision, be it on FreeBSD or Solaris.
The convenience is obvious. If you envision modern hardware, such as
SAS, the other question is what is reliability of corresponding driver,
here also a question on mailing lists may prove useful.



--

Michel TALON

From: Rob Kouwenberg on
Erick Engelke <erick(a)engmail.uwaterloo.ca> wrote:
> My team is looking at upgrading our FreeBSD fileserver, we've been
> extrememly happy with FreeBSD, the fileserver serves:
>
> umpteen NFS clients (all FreeBSD)
> 2,000 CIFS clients with Samba
> 9,000 homedirectories on about 4 TB of RAID disks on InforTrend RAID
> chassies, 2GB battery backed up RAM on RAID chassis
> with user quotas
> Our present box is dated. It's FreeBSD 6.2, dual processor, 4GB RAM
> running 32 bit OS. We will find some money to upgrade it.

what do e.g. w, top, lsof, vmstat 10 and netstat -an give on an hourly
basis ? Know you current load before heading of to the slaughterhouse !

For getting bigger capacity disks, on 1 standard sata2 port you can get
about 40Tb. 16 port sata2 cards, with 5 in the wintel box, would give
you 3.2Pb on the same hardware. Would that suffice ?

> We started on FreeBSD 2.0, so we've seen a lot of progress.

Heck, still running 2.2.5 here, both on a 4 watts server and in a vmware
imgae. It's getting about time to upgrade... :) Tough luck that BSD/OS
went south and I still can't make it vmized ...
From: Michael Sierchio on
Erick Engelke wrote:

> all 9,000 homedirectories on 16 TB RAID disks growable to 92 TB
>
> ZFS looks nice, except that to support 9,000 user accounts, we believe
> we would have to have 9,000 mount points. That's how the Sun system
> would work.

Doesn't have to -- it's conventional to make each user's home dir
a filesystem, but not at all required. It makes quota enforcement
easier, but you could do the old-fashioned /export/home/t/th/thisuser approach,
and have 30 mount points, or fewer, or more.

If you do go the ZFS approach, don't bother with RAID hardware at all --
JBOD and raidz1 or raidz2.

> Sun's Storage 7000 series family uses an 18 GB solid state hard disk
> which buffers the ZFS writes and is supposed to really speed up the
> system. Can I do the same with FreeBSD.

I'm skeptical about a SSD speeding up writes -- ZFS's copy-on-write
semantics have numerous advantages for conventional disk arrays, and
eliminate the write hole on conventional RAID5.

> In the end, we may purchase one of these brand name systems, but I want
> to give FreeBSD a fighting chance because it has been rock solid for all
> our server systems. We started on FreeBSD 2.0, so we've seen a lot of
> progress.

FreeBSD 8 would be a nice platform. I don't know that SATA2 disks would
be that much slower than SAS, and 2TB SATA2 drives are absurdly cheap. The
thing you run out of soonest is available bays to put disks in. Do you
factor power consumption and cooling requirements in your cost model?

- M
From: Erick Engelke on

On Tue, 5 Jan 2010, Michael Sierchio wrote:
> Erick Engelke wrote:
>
>> all 9,000 homedirectories on 16 TB RAID disks growable to 92 TB

Thanks to all who replied.

We've been monitorring our old system for quite a while. The load levels
are rarely above four, but at peak times the UFS hardware RAIDed SCSI
disks are nearing 100% utilization for sustained periods.

If we were to just upgrade our fileserver to FreeBSD 8 + UFS, we'd get
multicores, so our CPU load is not a problem. Right now our hardware has
us stuck at 4GB, but with new hardware we'd go for 64 GB or more. We
won't be cheap.

>> ZFS looks nice, except that to support 9,000 user accounts, we believe
>> we would have to have 9,000 mount points. That's how the Sun system
>> would work.
>
> Doesn't have to -- it's conventional to make each user's home dir
> a filesystem, but not at all required. It makes quota enforcement
> easier, but you could do the old-fashioned /export/home/t/th/thisuser approach,
> and have 30 mount points, or fewer, or more.

I thought I needed separate mount points to enforce quota at a per-user
basis.

My question is, will FreeBSD be okay with 9,0000 mount points. I know ZFS
requires a fair bit of memory, and I don't know if there is some sort of
formula for estimating how much memory for a given number of ZFS mount
points.

> If you do go the ZFS approach, don't bother with RAID hardware at all --
> JBOD and raidz1 or raidz2.

Right.

>
>> Sun's Storage 7000 series family uses an 18 GB solid state hard disk
>> which buffers the ZFS writes and is supposed to really speed up the
>> system. Can I do the same with FreeBSD.
>
> I'm skeptical about a SSD speeding up writes -- ZFS's copy-on-write
> semantics have numerous advantages for conventional disk arrays, and
> eliminate the write hole on conventional RAID5.

Interesting.

> FreeBSD 8 would be a nice platform. I don't know that SATA2 disks would
> be that much slower than SAS, and 2TB SATA2 drives are absurdly cheap. The
> thing you run out of soonest is available bays to put disks in. Do you
> factor power consumption and cooling requirements in your cost model?

No, we have ample cooling, power and space.

Thanks,

Erick