A computer components & hardware forum. HardwareBanter

If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below.

Go Back   Home » HardwareBanter forum » General Hardware & Peripherals » Storage (alternative)
Site Map Home Register Authors List Search Today's Posts Mark Forums Read Web Partners

RAID 5 vs. single SCSI drive?



 
 
Thread Tools Display Modes
  #1  
Old December 9th 03, 05:41 AM
Carlos Moreno
external usenet poster
 
Posts: n/a
Default RAID 5 vs. single SCSI drive?


Hi,

I'm trying to decide between these two options, for a server
running Linux (RedHat 9, most likely), which will be used as
a database server, running PostgreSQL (version 7.4, most
likely).

We are talking about peak load in the order of 20 or 30 inserts
per second, with clients connecting from three different
places.

The machine is a dual Athlon 2GHz with 1GB of memory.

Now, the big question is: what hard disk configuration should
I choose for this server?

I've been always reluctant to use SCSI; but I keep reading
that for server use, it's almost like it's not even a question,
it must be SCSI, period, by definition.

Considering the option of 3 IDE drives with RAID 5, how do
things compare? 3 IDE's with RAID 5 are a bit more expensive
than a single SCSI (we're talking about renting a dedicated
server, so we're talking monthly fees).

For this kind of use, should I be expecting better performance
from the IDE-based RAID 5 configuration?

What about two or three SCSI drives without RAID? Will this
be better than the IDE RAID 5 option? (2 SCSI drives of
36G each cost almost the same (a little bit more) than the
three IDE's with RAID5 configuration).

Thanks for any comments,

Carlos
--

  #2  
Old December 9th 03, 08:58 AM
Rod Speed
external usenet poster
 
Posts: n/a
Default


Carlos Moreno wrote
in message . ..

I'm trying to decide between these two options, for a server
running Linux (RedHat 9, most likely), which will be used as
a database server, running PostgreSQL (version 7.4, most
likely).


We are talking about peak load in the order of 20 or 30 inserts
per second, with clients connecting from three different places.


The machine is a dual Athlon 2GHz with 1GB of memory.


Now, the big question is: what hard disk
configuration should I choose for this server?


I've been always reluctant to use SCSI; but I keep
reading that for server use, it's almost like it's not
even a question, it must be SCSI, period, by definition.


Thats just plain wrong. The world's moved on.

Considering the option of 3 IDE drives with RAID 5,
how do things compare? 3 IDE's with RAID 5 are a
bit more expensive than a single SCSI


And MUCH more reiable.

(we're talking about renting a dedicated
server, so we're talking monthly fees).


For this kind of use, should I be expecting better
performance from the IDE-based RAID 5 configuration?


The reason for RAID5 aint performance, its reliability.

What about two or three SCSI drives without RAID?
Will this be better than the IDE RAID 5 option?


Not for reliability.

(2 SCSI drives of 36G each cost almost the same (a little
bit more) than the three IDE's with RAID5 configuration).


There's more involved than cost.


  #3  
Old December 9th 03, 02:58 PM
Carlos Moreno
external usenet poster
 
Posts: n/a
Default

Rod Speed wrote:

I've been always reluctant to use SCSI; but I keep
reading that for server use, it's almost like it's not
even a question, it must be SCSI, period, by definition.


Thats just plain wrong. The world's moved on.


Glad to hear that! :-)

However, I can't find references where SCSI is not a
definite winner when we're talking about servers. Am I
looking at the wrong place? Or are those comparisons
single SCSI vs. single IDE?

For this kind of use, should I be expecting better
performance from the IDE-based RAID 5 configuration?


The reason for RAID5 aint performance, its reliability.


Oh... I thought it was both. I was under the impression
that RAID 0 improves performance (half the data goes to
one drive, half to the other one, simultaneously), RAID 1
for reliability (same data goes to or comes from both
drives), and RAID 5 for both performance and reliability.
Am I mistaken on this? What would be the advantage of
RAID 5 over RAID 1 then?

(2 SCSI drives of 36G each cost almost the same (a little
bit more) than the three IDE's with RAID5 configuration).


There's more involved than cost.


Yes, of course. I mentioned the cost as a way of
establishing that "everything else is almost equal", so
that I would now decide purely based on technical merits.

That is, if two options cost the same, then a minor
advantage in one of them would make you decide. If
option A is twice as expensive, then you probably would
not go for it just for a 5% increase on performance
(well, it obviously depends on the application, but
you get the idea...)

Thanks!

Carlos
--

  #4  
Old December 9th 03, 03:57 PM
Will Dormann
external usenet poster
 
Posts: n/a
Default

Carlos Moreno wrote:
Considering the option of 3 IDE drives with RAID 5, how do
things compare? 3 IDE's with RAID 5 are a bit more expensive
than a single SCSI (we're talking about renting a dedicated
server, so we're talking monthly fees).

For this kind of use, should I be expecting better performance
from the IDE-based RAID 5 configuration?


Unless you have a fancy (expensive) RAID controller, the write
performance will be slower with RAID5 than a single drive. Though I
don't know for sure, I can't see anybody making a high-end IDE RAID5
controller. ("high-end" and "IDE RAID" together seems a bit of an
oxymoron).

If you're on a budget, a simple mirror will give you better write
performance than a 3-drive RAID 5 array. If you need high performance
+ redundancy, consider a RAID 1+0 array. (or a 0+1 array, if the
former isn't available)


-WD


-WD





-----= Posted via Newsfeeds.Com, Uncensored Usenet News =-----
http://www.newsfeeds.com - The #1 Newsgroup Service in the World!
-----== Over 100,000 Newsgroups - 19 Different Servers! =-----
  #5  
Old December 9th 03, 06:05 PM
Rod Speed
external usenet poster
 
Posts: n/a
Default


Carlos Moreno wrote in
message . ..
Rod Speed wrote


I've been always reluctant to use SCSI; but I keep
reading that for server use, it's almost like it's not
even a question, it must be SCSI, period, by definition.


Thats just plain wrong. The world's moved on.


Glad to hear that! :-)


However, I can't find references where SCSI is not
a definite winner when we're talking about servers.


Winner in what respect ? Clearly on price
alone, thats obvious from any pricelist.

Am I looking at the wrong place?


Its not something that gets a hell
of a lot of coverage on the web.

Or are those comparisons single SCSI vs. single IDE?


Certainly quite a few like with storagereview they are.

For this kind of use, should I be expecting better
performance from the IDE-based RAID 5 configuration?


The reason for RAID5 aint performance, its reliability.


Oh... I thought it was both.


Nope, RAID5 doesnt blow your sox off performance
wise, its primarily about reliability, maximum up time.

I was under the impression that RAID 0 improves
performance (half the data goes to one drive, half
to the other one, simultaneously),


Thats the theory, anyway. The reality can be
surprisingly dissapointing with modern hard drives.

Its mainly useful for stuff like raw video capture when
the speed of individual drives is just a bit too inadequate.

RAID 1 for reliability (same data goes to or comes from both
drives), and RAID 5 for both performance and reliability.


Nar. RAID5 splits the data over the drives in a more complex
way and uses proper ECCs to allow loss of one of the set
of drives, it uses more than 2 drives. There is a real performance
penalty with the calculation of the ECCs unless thats done in
hardware. The main advantage is that individual drive failure
has no real visible impact at all, so you get maximum availability.

Am I mistaken on this?


Yes.

What would be the advantage of RAID 5 over RAID 1 then?


Maximum availability. With a performance and purchase price penalty.

And there is a lot more than just 5 and 1 for reliability.

(2 SCSI drives of 36G each cost almost the same (a little
bit more) than the three IDE's with RAID5 configuration).


There's more involved than cost.


Yes, of course. I mentioned the cost as a way of
establishing that "everything else is almost equal", so
that I would now decide purely based on technical merits.


Trouble is that the difference in price can be ten times
between say a single 250GB IDE and a SCSI RAID5 array,
with very similar performance. Five times when comparing
two of those 250GB IDEs with the SCSI RAID5 array.

That is, if two options cost the same,


They never do with SCSI. Its always more expensive,
sometimes much more expensive like with the example above.

then a minor advantage in one of them would make you decide.


Sure, but thats not the real world.

If option A is twice as expensive, then you probably would
not go for it just for a 5% increase on performance


Yep, and thats the main reason that that particular
application is no longer SCSI without question.

(well, it obviously depends on the application, but you get the idea...)


Yep. But the reality is that the SCSI route is always substantially
more expensive, SCSI RAID5 doesnt necessarily give you any
better performance over modern IDE drives.

The most expensive route does bring with it some real advantages,
but thats mostly with a proper integrated hotswap package that
allows you to carry on regardless with a single drive failure and
swap the failed drive at your convenience. At a very high price.


  #6  
Old December 9th 03, 08:56 PM
Folkert Rienstra
external usenet poster
 
Posts: n/a
Default



"Carlos Moreno" wrote in message
Rod Speed wrote:

I've been always reluctant to use SCSI; but I keep
reading that for server use, it's almost like it's not
even a question, it must be SCSI, period, by definition.


Thats just plain wrong. The world's moved on.


Glad to hear that! :-)

However, I can't find references where SCSI is not a
definite winner when we're talking about servers. Am I
looking at the wrong place? Or are those comparisons
single SCSI vs. single IDE?

For this kind of use, should I be expecting better
performance from the IDE-based RAID 5 configuration?


The reason for RAID5 aint performance, its reliability.


Oh... I thought it was both.
I was under the impression
that RAID 0 improves performance (half the data goes to
one drive, half to the other one, simultaneously), RAID 1
for reliability (same data goes to or comes from both
drives), and RAID 5 for both performance and reliability.
Am I mistaken on this?


Well, that IS the theory.
Unfortunately, practice doesn't seem to know about that.

What would be the advantage of RAID 5 over RAID 1 then?

(2 SCSI drives of 36G each cost almost the same (a little
bit more) than the three IDE's with RAID5 configuration).


There's more involved than cost.


Yes, of course. I mentioned the cost as a way of
establishing that "everything else is almost equal", so
that I would now decide purely based on technical merits.

That is, if two options cost the same, then a minor
advantage in one of them would make you decide. If
option A is twice as expensive, then you probably would
not go for it just for a 5% increase on performance
(well, it obviously depends on the application, but
you get the idea...)

Thanks!

Carlos

  #7  
Old December 9th 03, 10:35 PM
J.Clarke
external usenet poster
 
Posts: n/a
Default

On Wed, 10 Dec 2003 21:46:22 +1100
"V@H" wrote:

software RAID-5 eats up CPU heaps...


However with a dual 2 GHz Athlon setup he's got heaps of CPU available.

"Carlos Moreno" wrote in message
. ..

Hi,

I'm trying to decide between these two options, for a server
running Linux (RedHat 9, most likely), which will be used as
a database server, running PostgreSQL (version 7.4, most
likely).

We are talking about peak load in the order of 20 or 30 inserts
per second, with clients connecting from three different
places.

The machine is a dual Athlon 2GHz with 1GB of memory.

Now, the big question is: what hard disk configuration should
I choose for this server?

I've been always reluctant to use SCSI; but I keep reading
that for server use, it's almost like it's not even a question,
it must be SCSI, period, by definition.

Considering the option of 3 IDE drives with RAID 5, how do
things compare? 3 IDE's with RAID 5 are a bit more expensive
than a single SCSI (we're talking about renting a dedicated
server, so we're talking monthly fees).

For this kind of use, should I be expecting better performance
from the IDE-based RAID 5 configuration?

What about two or three SCSI drives without RAID? Will this
be better than the IDE RAID 5 option? (2 SCSI drives of
36G each cost almost the same (a little bit more) than the
three IDE's with RAID5 configuration).

Thanks for any comments,

Carlos
--





--
--
--John
Reply to jclarke at ae tee tee global dot net
(was jclarke at eye bee em dot net)
  #8  
Old December 9th 03, 10:42 PM
Rita_A_Berkowitz
external usenet poster
 
Posts: n/a
Default



Though I don't know for sure, I can't see anybody making a high-end IDE

RAID5
controller. ("high-end" and "IDE RAID" together seems a bit of an
oxymoron).


Will, you are 99.9% correct with the exception that you forgot to add that
there are absolutely no "high-end" IDE-RAID controllers on the market and
probably never will be. Go SCSI young man, Go SCSI and never look back.



Rita




  #9  
Old December 10th 03, 08:08 AM
Marc de Vries
external usenet poster
 
Posts: n/a
Default

On Tue, 9 Dec 2003 17:42:52 -0500, "Rita_A_Berkowitz"
wrote:



Though I don't know for sure, I can't see anybody making a high-end IDE

RAID5
controller. ("high-end" and "IDE RAID" together seems a bit of an
oxymoron).


Will, you are 99.9% correct with the exception that you forgot to add that
there are absolutely no "high-end" IDE-RAID controllers on the market and
probably never will be. Go SCSI young man, Go SCSI and never look back.


You forgot to add that SCSI RAID5 controllers are also SLOW when
writing. ALL Raid5 solutions are slow when writing.

XOR calculations give a lot of overhead. BOTH on IDE and SCSI Raid5
controllers. The calculations are exactly the same. The processors
that do the calculations are also the same.
Thus the overhead that causes the slowdown in RAID5 is also the same.

There are reasons to choose SCSI, but this is NOT one of them.

Marc
  #10  
Old December 10th 03, 09:39 AM
Marc de Vries
external usenet poster
 
Posts: n/a
Default

On Tue, 09 Dec 2003 00:41:12 -0500, Carlos Moreno
wrote:


Hi,

I'm trying to decide between these two options, for a server
running Linux (RedHat 9, most likely), which will be used as
a database server, running PostgreSQL (version 7.4, most
likely).


I don't have experience with Linux and PostgreSQL, but I do have some
experience with MS SQL and lots with Exchange which are also databases
and should have similar hardware demands.

We are talking about peak load in the order of 20 or 30 inserts
per second, with clients connecting from three different
places.

The machine is a dual Athlon 2GHz with 1GB of memory.

Now, the big question is: what hard disk configuration should
I choose for this server?


The first question is if the harddiks will be a bottleneck at all. Do
you have experience with other machines running a similar
configuration?

Often a lot of memory will make make load on the harddisk a lot lower.
But this depends lot on the size of the database and the way it is
used. If the database easily fits in memory the harddisk is not used
all that much.

I've been always reluctant to use SCSI; but I keep reading
that for server use, it's almost like it's not even a question,
it must be SCSI, period, by definition.


IDE was not suited for servers in the past. This is the main reason
SCSI is the defacto standard in servers. The server market is very
conservative and people more easily spend the money of the company
then their own money, which is another reason SCSI has stayed the
standard.

Still there are good reasons to choose SCSI is you are willing to pay
premium prices.

- SCSI has much longer cables then IDE. Which can be a pratical issue
in larger servers. Especially when you have 19" racks with external
storage enclosures. (SATA2 will change all that)
- Until recently you only had Raid controllers for SCSI. Which is
considered necessary for servers also. There are good RAID controllers
for IDE for small servers nowadays. In ranges for 4, 8 or even 12
disks. But if you want to have more disks on your array controller you
still need SCSI.

- High end SCSI array controllers (which are VERY expensive) have more
reliability features then IDE controllers. For example external power
supplies in case the mobo fails, and onboard batteries for the ram on
the array controller. All designed so that you can have write back
cache on the controller, and not loose data when power fails in your
server.

- Because SCSI is de defacto standard for servers, manufacturers have
made the more reliable harddisks mainly for SCSI. There is no
technical reason why they couldn't have made IDE harddisks with the
same MTBF, but there simply wasn't a market for it.
This has changed recently with the WD Raptor 10.000 rpm IDE harddisk
and the Maxtor Maxline II 7200 rpm harddisk. The WD Raptor is pretty
expensive though. (but SCSI is still more expensive)
Then again, reliability of harddisk is not all that important anymore
when you use Raid arrays. Even when a IDE disks fails twice as fast,
it is still a lot cheaper.

But I have the impression that these kinds of features of SCSI are not
really what you are interested in.

Still the most important reason to choose SCSI is because of the very
low access times. This again is not caused by the SCSI protocol
itself, but by the high rpm values. A 15000 rpm SCSI has a much better
access time then 10.000 rpm SCSI or 7200 rpm IDE disks.

This access time is important for Random read/write operations.
Which doesn't happen a lot on your desktop, or even on fileservers and
webservers.
SCSI drivers have simple cache management which is optimized for
randrom read/write operations. IDE drivers cache management is
optimized for typical desktop use.
This is why 7200 rpm IDE drives will regularly beat 10.000 rpm SCSI
(and sometimes even 15.000 rpm SCSI) drives in desktop situations.

But a database often has lots of random read/write operations.
This is what makes a database the best case scenario for a SCSI disk.
(I'm talking about comparing single disks here without looking at the
huge price difference)

Considering the option of 3 IDE drives with RAID 5, how do
things compare? 3 IDE's with RAID 5 are a bit more expensive
than a single SCSI (we're talking about renting a dedicated
server, so we're talking monthly fees).

For this kind of use, should I be expecting better performance
from the IDE-based RAID 5 configuration?


It depends.
RAID5 is mainly designed to have affordable redundancy.
It is fast when you read data, but the XOR calculations for RAID5
makes it slow for writing data.
This is a feature of RAID5 itself, and thus applies to both IDE and
SCSI Raid5 controllers.

If you want ultimate performance, and also redundancy then you should
use RAID10. (or RAID0+1)

You would need 4 disks of which two are used for data, instead of 3 of
which two are used with Raid5, but is is a lot faster when writing and
usually also a bit faster when reading. The controller itself can be a
lot cheaper.
4 IDE disks with a Raid10 controller are probably cheaper then 3 IDE
disks with a Raid5 controller.

4 IDE's in Raid10 will be faster than a single SCSI in most situations
and have the added benefit of redundancy. There have been some review
sites that have tested this, but I can't remember them just now.
It has probably been anandtech and/or xbitlabs.

What about two or three SCSI drives without RAID? Will this
be better than the IDE RAID 5 option? (2 SCSI drives of
36G each cost almost the same (a little bit more) than the
three IDE's with RAID5 configuration).


Regardless of SCSI or IDE you should also consider if you want RAID or
want multiple seperate disk volumes.

In lots of databases you have a seperate disk for data and a seperate
disk for logfiles. When a lot of data is written to the logs files, it
gives a huge performance boast to place those logfiles on a seperate
disk, so that the database disk can do other tasks at the same time.

Being able to do that may well be the most important reason to choose
multiple IDE disks over a single SCSI disk.

Of course you might also opt for a single IDE disk for the logfiles
and a IDE RAID0+1 volume for the database files.

For smaller servers there will be lots of situations where you can
configure a server with IDE that will perform just as well as SCSI for
a lower price. Or configure a server that will perform better then
SCSI for the same price.

But when the database gets more demanding there will come a point
where a single SCSI disk, or a few IDE disks in a Raid volume can't
deliver the performance your database wants.
At that time you have to consider buying SCSI raid controllers.

So, as you see, SCSI still has it's place in servers, but it's not
that easy anymore to determine when it is the best solution.

I hope this will give you some more ideas on how to configure your
server.

Also you might want to check some reviews:
http://www.tech-report.com/reviews/2...o/index.x?pg=1
http://www.tech-report.com/reviews/2...k/index.x?pg=1
http://www6.tomshardware.com/storage...114/index.html

And there are lots more that are interesting.

Marc
 




Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

vB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Forum Jump

Similar Threads
Thread Thread Starter Forum Replies Last Post
Question about transfer speeds between HDs, and DMA mode ZigZag Master Overclocking AMD Processors 25 July 25th 04 09:56 PM
P4P800DLX from non-raid to raid Splitskull Asus Motherboards 2 June 2nd 04 10:51 AM
Gigabyte GA-8KNXP and Promise SX4000 RAID Controller Old Dude Gigabyte Motherboards 4 November 12th 03 07:26 PM
help with motherboard choice S.Boardman General 30 October 20th 03 10:23 PM
GA-8KNXP, how to configure BIOS for SATA? John Ward Gigabyte Motherboards 20 October 6th 03 07:42 AM


All times are GMT +1. The time now is 06:35 PM.


Powered by vBulletin® Version 3.6.4
Copyright ©2000 - 2024, Jelsoft Enterprises Ltd.
Copyright ©2004-2024 HardwareBanter.
The comments are property of their posters.