[Asterisk-Users] RAID affecting X100P performance...

Scott Laird scott at sigkill.org
Wed Jul 21 20:47:53 MST 2004


On Jul 21, 2004, at 7:01 PM, Steven Critchfield wrote:
> BTW, my raid card on my Dell 2450 had this output
> nash5:/home/critch# hdparm -tT /dev/sda
>
> /dev/sda:
>  Timing buffer-cache reads:   128 MB in  0.61 seconds =209.84 MB/sec
>  Timing buffered disk reads:  64 MB in  2.52 seconds = 25.40 MB/sec
>
> and I don't have xeons, let alone another 10% to give up to deal with
> the drives. I doubt you are getting 200mb/s real data movements over 
> IDE
> and software raid.

I don't have my notes in front of me, but I was seeing almost 200 
MB/sec reads through bonnie with a multi-GB sample size the last time I 
tried all of this.  Admittedly, it's been a couple years, and I was 
pushing a *lot* of drives (16, in the last incarnation).  Heck, I was 
getting 70-80 MB/sec with software RAID-5 over 5 SCSI drives in 1999.  
That's quite a bit better then the 25 MB/sec you're seeing.  Like I 
said, I haven't had good experiences with any low-cost SCSI RAID 
cards--they all run out of CPU horsepower before your drives run out of 
throughput or your host CPU gets loaded.

Not that it really matters most of the time--most real-world jobs are 
limited by random I/O or cost, not streaming disk throughput.  
Particularly if networks are involved.

> Raid5 compared to mirror should be easier on raid5. Write 3k data over 
> 4
> drive raid5 array and you only write 4k total and the xor is cheap and
> easy to do. 3k over a mirror means 6k is written. So if you do a 1/3rd
> more slow task work, which is more of a cpu hit?

Do the math again.  With RAID 5, writting a small amount of data with a 
cold cache will cost 1 read from each drive plus 1 write to two drives. 
  For 5 drives, that's a total of 7 I/Os, which is way worse then the 2 
you'd need for RAID 1.  As the writes get bigger then the RAID block 
size, reads drop to 0 and you're only paying a slight overhead for RAID 
5, but small writes really suck with RAID 5.

>> For this (and a number of other reasons), you're best off avoiding 
>> RAID
>> 5 if you care about random I/O performance.  It can be made to go 
>> fast,
>> but you'll need to throw a lot of cash at it.  The same amount of cash
>> will frequently get you better performance with RAID 0+1 (or 1+0,
>> depending on how you look at things).
>
> Truethfully, if you want cheap, forget raid and just keep a cold spare
> drive. You probably are going to be down for a while anyways.

Personally, if the server doesn't have any critical data (like DNS or a 
non-VM Asterisk server), it's easier to keep a live spare system, and 
plan on failing over as quickly as possible.  With a bit of design 
work, most services that don't require a lot of state are happiest this 
way, and it lets you get away with dirt cheap servers.

If you have something like a busy file server, database, mail server, 
voicemail server, or anything with heavyweight sessions, then it's time 
to look at RAID.

> oddly enough, there isn't much if any difference these days at the
> physical level. It is just the interface and the set of specs on the
> interface. SCSI drives usually will give you warning of their problems.

IDE with SMART will warn, too.  The big difference is tagged 
queueing--almost all SCSI drives support it, and almost no IDE drives 
do.  It lets the controller hand the drive a long list of I/Os to 
perform, and the drive can optimize the order of the reads to best fit 
rotational and seek latency.  In some cases, this will give you 2x the 
performance.  That, plus the fact that you can get 15k RPM SCSI drives 
gives SCSI a substantial per-drive advantage and an unbeatable latency 
advantage.  For me, most of the time, IDE's cost advantage wins out, 
though.

> If it is in a business setup, it should be hotswappable if you bother
> with raid at all.

Hot swap is nice, when it works.  For two drives, if you can afford 
scheduled downtime and you don't have a lot of systems to babysit, then 
it's probably overkill.  At my last job, I rolled out hundreds of 
1-drive hot-swap systems, just because the local help couldn't be 
trusted to open up a case and swap drives without breaking things.  It 
all depends on the environment.  In most of my environments today, I 
wouldn't bother with SCSI unless I was planning on serious amounts of 
random disk I/O.


Scott




More information about the asterisk-users mailing list