If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below. |
|
|
Thread Tools | Display Modes |
#1
|
|||
|
|||
RAID 5 vs. single SCSI drive?
Hi, I'm trying to decide between these two options, for a server running Linux (RedHat 9, most likely), which will be used as a database server, running PostgreSQL (version 7.4, most likely). We are talking about peak load in the order of 20 or 30 inserts per second, with clients connecting from three different places. The machine is a dual Athlon 2GHz with 1GB of memory. Now, the big question is: what hard disk configuration should I choose for this server? I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Considering the option of 3 IDE drives with RAID 5, how do things compare? 3 IDE's with RAID 5 are a bit more expensive than a single SCSI (we're talking about renting a dedicated server, so we're talking monthly fees). For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? What about two or three SCSI drives without RAID? Will this be better than the IDE RAID 5 option? (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). Thanks for any comments, Carlos -- |
#2
|
|||
|
|||
Carlos Moreno wrote in message . .. I'm trying to decide between these two options, for a server running Linux (RedHat 9, most likely), which will be used as a database server, running PostgreSQL (version 7.4, most likely). We are talking about peak load in the order of 20 or 30 inserts per second, with clients connecting from three different places. The machine is a dual Athlon 2GHz with 1GB of memory. Now, the big question is: what hard disk configuration should I choose for this server? I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Thats just plain wrong. The world's moved on. Considering the option of 3 IDE drives with RAID 5, how do things compare? 3 IDE's with RAID 5 are a bit more expensive than a single SCSI And MUCH more reiable. (we're talking about renting a dedicated server, so we're talking monthly fees). For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? The reason for RAID5 aint performance, its reliability. What about two or three SCSI drives without RAID? Will this be better than the IDE RAID 5 option? Not for reliability. (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). There's more involved than cost. |
#3
|
|||
|
|||
Rod Speed wrote:
I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Thats just plain wrong. The world's moved on. Glad to hear that! :-) However, I can't find references where SCSI is not a definite winner when we're talking about servers. Am I looking at the wrong place? Or are those comparisons single SCSI vs. single IDE? For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? The reason for RAID5 aint performance, its reliability. Oh... I thought it was both. I was under the impression that RAID 0 improves performance (half the data goes to one drive, half to the other one, simultaneously), RAID 1 for reliability (same data goes to or comes from both drives), and RAID 5 for both performance and reliability. Am I mistaken on this? What would be the advantage of RAID 5 over RAID 1 then? (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). There's more involved than cost. Yes, of course. I mentioned the cost as a way of establishing that "everything else is almost equal", so that I would now decide purely based on technical merits. That is, if two options cost the same, then a minor advantage in one of them would make you decide. If option A is twice as expensive, then you probably would not go for it just for a 5% increase on performance (well, it obviously depends on the application, but you get the idea...) Thanks! Carlos -- |
#4
|
|||
|
|||
Carlos Moreno wrote:
Considering the option of 3 IDE drives with RAID 5, how do things compare? 3 IDE's with RAID 5 are a bit more expensive than a single SCSI (we're talking about renting a dedicated server, so we're talking monthly fees). For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? Unless you have a fancy (expensive) RAID controller, the write performance will be slower with RAID5 than a single drive. Though I don't know for sure, I can't see anybody making a high-end IDE RAID5 controller. ("high-end" and "IDE RAID" together seems a bit of an oxymoron). If you're on a budget, a simple mirror will give you better write performance than a 3-drive RAID 5 array. If you need high performance + redundancy, consider a RAID 1+0 array. (or a 0+1 array, if the former isn't available) -WD -WD -----= Posted via Newsfeeds.Com, Uncensored Usenet News =----- http://www.newsfeeds.com - The #1 Newsgroup Service in the World! -----== Over 100,000 Newsgroups - 19 Different Servers! =----- |
#5
|
|||
|
|||
Carlos Moreno wrote in message . .. Rod Speed wrote I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Thats just plain wrong. The world's moved on. Glad to hear that! :-) However, I can't find references where SCSI is not a definite winner when we're talking about servers. Winner in what respect ? Clearly on price alone, thats obvious from any pricelist. Am I looking at the wrong place? Its not something that gets a hell of a lot of coverage on the web. Or are those comparisons single SCSI vs. single IDE? Certainly quite a few like with storagereview they are. For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? The reason for RAID5 aint performance, its reliability. Oh... I thought it was both. Nope, RAID5 doesnt blow your sox off performance wise, its primarily about reliability, maximum up time. I was under the impression that RAID 0 improves performance (half the data goes to one drive, half to the other one, simultaneously), Thats the theory, anyway. The reality can be surprisingly dissapointing with modern hard drives. Its mainly useful for stuff like raw video capture when the speed of individual drives is just a bit too inadequate. RAID 1 for reliability (same data goes to or comes from both drives), and RAID 5 for both performance and reliability. Nar. RAID5 splits the data over the drives in a more complex way and uses proper ECCs to allow loss of one of the set of drives, it uses more than 2 drives. There is a real performance penalty with the calculation of the ECCs unless thats done in hardware. The main advantage is that individual drive failure has no real visible impact at all, so you get maximum availability. Am I mistaken on this? Yes. What would be the advantage of RAID 5 over RAID 1 then? Maximum availability. With a performance and purchase price penalty. And there is a lot more than just 5 and 1 for reliability. (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). There's more involved than cost. Yes, of course. I mentioned the cost as a way of establishing that "everything else is almost equal", so that I would now decide purely based on technical merits. Trouble is that the difference in price can be ten times between say a single 250GB IDE and a SCSI RAID5 array, with very similar performance. Five times when comparing two of those 250GB IDEs with the SCSI RAID5 array. That is, if two options cost the same, They never do with SCSI. Its always more expensive, sometimes much more expensive like with the example above. then a minor advantage in one of them would make you decide. Sure, but thats not the real world. If option A is twice as expensive, then you probably would not go for it just for a 5% increase on performance Yep, and thats the main reason that that particular application is no longer SCSI without question. (well, it obviously depends on the application, but you get the idea...) Yep. But the reality is that the SCSI route is always substantially more expensive, SCSI RAID5 doesnt necessarily give you any better performance over modern IDE drives. The most expensive route does bring with it some real advantages, but thats mostly with a proper integrated hotswap package that allows you to carry on regardless with a single drive failure and swap the failed drive at your convenience. At a very high price. |
#6
|
|||
|
|||
"Carlos Moreno" wrote in message Rod Speed wrote: I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Thats just plain wrong. The world's moved on. Glad to hear that! :-) However, I can't find references where SCSI is not a definite winner when we're talking about servers. Am I looking at the wrong place? Or are those comparisons single SCSI vs. single IDE? For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? The reason for RAID5 aint performance, its reliability. Oh... I thought it was both. I was under the impression that RAID 0 improves performance (half the data goes to one drive, half to the other one, simultaneously), RAID 1 for reliability (same data goes to or comes from both drives), and RAID 5 for both performance and reliability. Am I mistaken on this? Well, that IS the theory. Unfortunately, practice doesn't seem to know about that. What would be the advantage of RAID 5 over RAID 1 then? (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). There's more involved than cost. Yes, of course. I mentioned the cost as a way of establishing that "everything else is almost equal", so that I would now decide purely based on technical merits. That is, if two options cost the same, then a minor advantage in one of them would make you decide. If option A is twice as expensive, then you probably would not go for it just for a 5% increase on performance (well, it obviously depends on the application, but you get the idea...) Thanks! Carlos |
#7
|
|||
|
|||
On Wed, 10 Dec 2003 21:46:22 +1100
"V@H" wrote: software RAID-5 eats up CPU heaps... However with a dual 2 GHz Athlon setup he's got heaps of CPU available. "Carlos Moreno" wrote in message . .. Hi, I'm trying to decide between these two options, for a server running Linux (RedHat 9, most likely), which will be used as a database server, running PostgreSQL (version 7.4, most likely). We are talking about peak load in the order of 20 or 30 inserts per second, with clients connecting from three different places. The machine is a dual Athlon 2GHz with 1GB of memory. Now, the big question is: what hard disk configuration should I choose for this server? I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. Considering the option of 3 IDE drives with RAID 5, how do things compare? 3 IDE's with RAID 5 are a bit more expensive than a single SCSI (we're talking about renting a dedicated server, so we're talking monthly fees). For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? What about two or three SCSI drives without RAID? Will this be better than the IDE RAID 5 option? (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). Thanks for any comments, Carlos -- -- -- --John Reply to jclarke at ae tee tee global dot net (was jclarke at eye bee em dot net) |
#8
|
|||
|
|||
Though I don't know for sure, I can't see anybody making a high-end IDE RAID5 controller. ("high-end" and "IDE RAID" together seems a bit of an oxymoron). Will, you are 99.9% correct with the exception that you forgot to add that there are absolutely no "high-end" IDE-RAID controllers on the market and probably never will be. Go SCSI young man, Go SCSI and never look back. Rita |
#9
|
|||
|
|||
On Tue, 9 Dec 2003 17:42:52 -0500, "Rita_A_Berkowitz"
wrote: Though I don't know for sure, I can't see anybody making a high-end IDE RAID5 controller. ("high-end" and "IDE RAID" together seems a bit of an oxymoron). Will, you are 99.9% correct with the exception that you forgot to add that there are absolutely no "high-end" IDE-RAID controllers on the market and probably never will be. Go SCSI young man, Go SCSI and never look back. You forgot to add that SCSI RAID5 controllers are also SLOW when writing. ALL Raid5 solutions are slow when writing. XOR calculations give a lot of overhead. BOTH on IDE and SCSI Raid5 controllers. The calculations are exactly the same. The processors that do the calculations are also the same. Thus the overhead that causes the slowdown in RAID5 is also the same. There are reasons to choose SCSI, but this is NOT one of them. Marc |
#10
|
|||
|
|||
On Tue, 09 Dec 2003 00:41:12 -0500, Carlos Moreno
wrote: Hi, I'm trying to decide between these two options, for a server running Linux (RedHat 9, most likely), which will be used as a database server, running PostgreSQL (version 7.4, most likely). I don't have experience with Linux and PostgreSQL, but I do have some experience with MS SQL and lots with Exchange which are also databases and should have similar hardware demands. We are talking about peak load in the order of 20 or 30 inserts per second, with clients connecting from three different places. The machine is a dual Athlon 2GHz with 1GB of memory. Now, the big question is: what hard disk configuration should I choose for this server? The first question is if the harddiks will be a bottleneck at all. Do you have experience with other machines running a similar configuration? Often a lot of memory will make make load on the harddisk a lot lower. But this depends lot on the size of the database and the way it is used. If the database easily fits in memory the harddisk is not used all that much. I've been always reluctant to use SCSI; but I keep reading that for server use, it's almost like it's not even a question, it must be SCSI, period, by definition. IDE was not suited for servers in the past. This is the main reason SCSI is the defacto standard in servers. The server market is very conservative and people more easily spend the money of the company then their own money, which is another reason SCSI has stayed the standard. Still there are good reasons to choose SCSI is you are willing to pay premium prices. - SCSI has much longer cables then IDE. Which can be a pratical issue in larger servers. Especially when you have 19" racks with external storage enclosures. (SATA2 will change all that) - Until recently you only had Raid controllers for SCSI. Which is considered necessary for servers also. There are good RAID controllers for IDE for small servers nowadays. In ranges for 4, 8 or even 12 disks. But if you want to have more disks on your array controller you still need SCSI. - High end SCSI array controllers (which are VERY expensive) have more reliability features then IDE controllers. For example external power supplies in case the mobo fails, and onboard batteries for the ram on the array controller. All designed so that you can have write back cache on the controller, and not loose data when power fails in your server. - Because SCSI is de defacto standard for servers, manufacturers have made the more reliable harddisks mainly for SCSI. There is no technical reason why they couldn't have made IDE harddisks with the same MTBF, but there simply wasn't a market for it. This has changed recently with the WD Raptor 10.000 rpm IDE harddisk and the Maxtor Maxline II 7200 rpm harddisk. The WD Raptor is pretty expensive though. (but SCSI is still more expensive) Then again, reliability of harddisk is not all that important anymore when you use Raid arrays. Even when a IDE disks fails twice as fast, it is still a lot cheaper. But I have the impression that these kinds of features of SCSI are not really what you are interested in. Still the most important reason to choose SCSI is because of the very low access times. This again is not caused by the SCSI protocol itself, but by the high rpm values. A 15000 rpm SCSI has a much better access time then 10.000 rpm SCSI or 7200 rpm IDE disks. This access time is important for Random read/write operations. Which doesn't happen a lot on your desktop, or even on fileservers and webservers. SCSI drivers have simple cache management which is optimized for randrom read/write operations. IDE drivers cache management is optimized for typical desktop use. This is why 7200 rpm IDE drives will regularly beat 10.000 rpm SCSI (and sometimes even 15.000 rpm SCSI) drives in desktop situations. But a database often has lots of random read/write operations. This is what makes a database the best case scenario for a SCSI disk. (I'm talking about comparing single disks here without looking at the huge price difference) Considering the option of 3 IDE drives with RAID 5, how do things compare? 3 IDE's with RAID 5 are a bit more expensive than a single SCSI (we're talking about renting a dedicated server, so we're talking monthly fees). For this kind of use, should I be expecting better performance from the IDE-based RAID 5 configuration? It depends. RAID5 is mainly designed to have affordable redundancy. It is fast when you read data, but the XOR calculations for RAID5 makes it slow for writing data. This is a feature of RAID5 itself, and thus applies to both IDE and SCSI Raid5 controllers. If you want ultimate performance, and also redundancy then you should use RAID10. (or RAID0+1) You would need 4 disks of which two are used for data, instead of 3 of which two are used with Raid5, but is is a lot faster when writing and usually also a bit faster when reading. The controller itself can be a lot cheaper. 4 IDE disks with a Raid10 controller are probably cheaper then 3 IDE disks with a Raid5 controller. 4 IDE's in Raid10 will be faster than a single SCSI in most situations and have the added benefit of redundancy. There have been some review sites that have tested this, but I can't remember them just now. It has probably been anandtech and/or xbitlabs. What about two or three SCSI drives without RAID? Will this be better than the IDE RAID 5 option? (2 SCSI drives of 36G each cost almost the same (a little bit more) than the three IDE's with RAID5 configuration). Regardless of SCSI or IDE you should also consider if you want RAID or want multiple seperate disk volumes. In lots of databases you have a seperate disk for data and a seperate disk for logfiles. When a lot of data is written to the logs files, it gives a huge performance boast to place those logfiles on a seperate disk, so that the database disk can do other tasks at the same time. Being able to do that may well be the most important reason to choose multiple IDE disks over a single SCSI disk. Of course you might also opt for a single IDE disk for the logfiles and a IDE RAID0+1 volume for the database files. For smaller servers there will be lots of situations where you can configure a server with IDE that will perform just as well as SCSI for a lower price. Or configure a server that will perform better then SCSI for the same price. But when the database gets more demanding there will come a point where a single SCSI disk, or a few IDE disks in a Raid volume can't deliver the performance your database wants. At that time you have to consider buying SCSI raid controllers. So, as you see, SCSI still has it's place in servers, but it's not that easy anymore to determine when it is the best solution. I hope this will give you some more ideas on how to configure your server. Also you might want to check some reviews: http://www.tech-report.com/reviews/2...o/index.x?pg=1 http://www.tech-report.com/reviews/2...k/index.x?pg=1 http://www6.tomshardware.com/storage...114/index.html And there are lots more that are interesting. Marc |
Thread Tools | |
Display Modes | |
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Question about transfer speeds between HDs, and DMA mode | ZigZag Master | Overclocking AMD Processors | 25 | July 25th 04 09:56 PM |
P4P800DLX from non-raid to raid | Splitskull | Asus Motherboards | 2 | June 2nd 04 10:51 AM |
Gigabyte GA-8KNXP and Promise SX4000 RAID Controller | Old Dude | Gigabyte Motherboards | 4 | November 12th 03 07:26 PM |
help with motherboard choice | S.Boardman | General | 30 | October 20th 03 10:23 PM |
GA-8KNXP, how to configure BIOS for SATA? | John Ward | Gigabyte Motherboards | 20 | October 6th 03 07:42 AM |