Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature currently requires accessing the site using the built-in Safari browser.
Under Linux the Marvell 88SE6480 chip used in SuperMicro AOC-SASLP-MV8 works pretty well; but at least in BSD this controller sucks with continuous timeouts and other problems.
This I doubt very much. As far as I know, all Sun/Oracle storage servers that use ZFS are using HBA only. No one use expanders. Unless you can show links that confirms your statement, I will not accept your statement. I dont think your statement is correct, unfortunately.Sun sells large numbers of JBOD chassis that are based on SAS exapnders and explicitly support SATA disks in them.
Originally Posted by mikesm
Sun sells large numbers of JBOD chassis that are based on SAS exapnders and explicitly support SATA disks in them.
This I doubt very much.
...SNIP...
Updated by Jay Worthington 3 months ago
Richard Elling wrote:
SATA devices directly connected to SAS expanders is toxic.
In case someone cares, Solaris 11, released yesterday, fixes this "hardware"-problem :-D.
Now all we need is for Larry to fork over that promissed sourcecode so it can be fixed here as well
Regards,
Jay
Updated by Phillip Steinbachs 3 months ago
Jay Worthington wrote:
Richard Elling wrote:
SATA devices directly connected to SAS expanders is toxic.
In case someone cares, Solaris 11, released yesterday, fixes this "hardware"-problem :-D.
Now all we need is for Larry to fork over that promissed sourcecode so it can be fixed here as well
Thanks Jay for the followup. I'm glad to hear that Oracle devoted resources to resolving the problem and I hope that given enough time, a fix can be incorporated into Illumos and ultimately NexentaStor. It certainly presents a compelling reason in the interim to consider Solaris 11, even if it means giving more chump change to Larry.
-phillip
I claimed from the first day the Nexenta folks started posting on this that they were wrong - that it was not a fundamental fault of using SATA disks with SAS expanders - but that it indicated clearly some software fault and fragility of the ZFS or IO driver implementation. Of course, since they are part of the priesthood of ZFS they could not accept the idea that any fault could possibly be due to the ZFS implementation or the drivers. It would be sacrilege to even suggest that...even though no other use case for using SAS expanders with SATA drivers suffered from this fault.
At the time, my suggestion that this uncovered an instability in ZFS and/or the OS drivers was answered with the thought that ZFS was "just so good" that it uncovered faults with the hardware unseen by "inferior" applications like hardware RAIDs. They've gone on to re-post their "its toxic" thing over and over again here on [H] as well as on other forums.
It is a fact that zfs catches errors that no other filesystem/hw-raid does. It is a fact that zfs is much more sensitive and notices errors very quickly. It is a fact that people have used faulty hardware getting no errors reports, until they switched to zfs. These are facts.At the time, my suggestion that this uncovered an instability in ZFS and/or the OS drivers was answered with the thought that ZFS was "just so good" that it uncovered faults with the hardware unseen by "inferior" applications like hardware RAIDs. They've gone on to re-post their "its toxic" thing over and over again here on [H] as well as on other forums.
Woo. Strong words in here, and though I note they lack any data to back them up, I am ill-equipped to rebut on this topic, as I was not involved with any of the testing and analysis that led to the conclusion that SATA over SAS is bad. I trumpet it as far and wide as I can because I have seen the symptoms and outcome of the problem, and they are not something you want to experience in a production environment.
It is a fact that zfs is much more sensitive and notices errors very quickly. It is a fact that people have used faulty hardware getting no errors reports, until they switched to zfs. These are facts.
He's trumpeting a solution that works to a problem he's seen. Geez.You freely admit that you are not equipped with any data of your own, yet you trumpet the findings that you admit you don't actually understand because you have seen "symptoms" of "the problem". Even though the "symptoms" are equally well explained by a fault in MPIO (as has now been proven by Orcacle) and the fact that these symptoms do not afflict any other known use of SATA drives with SAS expanders - even in applications that are every bit ZFS's equal in total performance and stress on the drives.
Cute.
He's trumpeting a solution that works to a problem he's seen. Geez.
Even though the "symptoms" are equally well explained by a fault in MPIO (as has now been proven by Orcacle)
And because the wonderful folks at Nexenta (and @Nex7 above) can't accept that there might be a bug they blame the hardware.
My complaint is with the people like the last two posters who will post FUD as if it is fact - even when they don't even have the faintest understanding of what they are posting.
Also, I've been doing a bit more digging here. The bug they have fixed in MPIO was not specific to the use of expanders. It is a race condition that can occur after other drive signalling errors as well. Yes, the combination of SAS expanders and SATA disks did make it more likely and tickle the bug - but until this fix is propagated then Nexenta and other non-S11 ZFS implementations could still fault. With or without expanders. And as long as the folks at Nexenta keep their head in the sand over this they are putting their customer's data at risk.
Well said. It is good that you show some balance in this matter. If Oracle has fixed this problem, the better. However, I think we need more information on this, and confirmation, for instance by Garret Damore who blogged about this problem. Also, if SAS expanders + SATA disks are problematic combo right now, it should best be avoided until we have confirmed that it is an ok combo?These "facts" may be true. But that is not what is in play here. We are discussing a hardware configuration that works perfectly in all reported applications outside of ZFS that can cause a complete meltdown of a ZFS filesystem. No rational commercial user would support the use of SATA disks with SAS expanders if what is suggested were true. They demand systems that are robust and stable. ZFS is, in fact, very robust and stable in the face of a wide variety of fault conditions. It is shocking to me that any experienced, rational thinking person would ever suggest that ZFS is "more sensitive" in a way that would cause a complete meltdown and somehow be proud of ZFS for doing this. The entire intent of a platform like ZFS is to survive and be robust - which is does admirably - except in this case. And because the wonderful folks at Nexenta (and @Nex7 above) can't accept that there might be a bug they blame the hardware. Completely silly.
My dispute is not with ZFS. I think its wonderful. My complaint is with the people like the last two posters who will post FUD as if it is fact - even when they don't even have the faintest understanding of what they are posting.
The only thing I'd like Nex7 to answer is his contacts for SAS drive pricing. It seems as though the large capacity drives are drying up, I need to expand my storage solution and always preferred SAS drives, its just too expensive.
Also, I've been doing a bit more digging here. The bug they have fixed in MPIO was not specific to the use of expanders. It is a race condition that can occur after other drive signalling errors as well. Yes, the combination of SAS expanders and SATA disks did make it more likely and tickle the bug - but until this fix is propagated then Nexenta and other non-S11 ZFS implementations could still fault. With or without expanders. And as long as the folks at Nexenta keep their head in the sand over this they are putting their customer's data at risk.
I wish I could! I do not personally work in Sales or Sales Engineering so I don't get a lot of first-hand contacts in this area. I know that a few years ago at my last employer, we got 7200 RPM enterprise SAS high-capacity disks at a price difference from 7200 RPM enterprise SATA high-capacity disks that could only be described as 'completely negligible'. It would thus stand to reason that today, one could get them even closer to price-parity. However, the natural disasters curbing supply so heavily have broken a lot of that. I have heard that even some larger resellers have also had issues sourcing SAS disks at the moment. The supply just isn't there.![]()
I also want to respond to this on its own merits, because if there's a single case of FUD in this topic, it's this paragraph.
As I have mentioned previously, I possess no unique and total knowledge about what causes what I coin the 'SATA over SAS' problem. I have admitted to that, and admitted to my most likely suspects. That said, my year and a half of time spent in the Support department at Nexenta, probably one of the few or only companies in the entire world offering a storage appliance based on non-S11 ZFS that you can run on your own hardware (including SATA over SAS), makes me part of what is I believe a very small group of people who have real-world experience on a multitude of systems that have seen this problem. I have seen it often enough to recognize it almost immediately. I am intimately familiar with the SYMPTOMS and the potential fixes for this issue, it is only the root cause that I am not clear on (and let's be clear here -- I AM sure it involves SATA over SAS involving expanders, because that is always where it appears and so far changing that is always how it goes away).
Let me be perfectly clear. No equivocation. I have never seen, nor heard, of a single incident involving the symptoms (or even most of the symptoms) I equate with the SATA over SAS bug in an all-SAS environment. Not on Nexenta, nor on OpenIndiana, SmartOS, or any other illumos-derived distribution. Not one time. Not one.
I have further witnessed multiple examples where the issue was occurring and the removal of SATA component OR the removal of all expanders caused an immediate partial or total relief. I would never suggest to potential customers to go all-SAS as opposed to SATA if I'd seen the problem in both environments, and to suggest otherwise is, to be honest, a little insulting.
Based on the fact that in the 1000's of systems I've been in the position to become aware of should they have ever seen this, and the multitude of systems incorporating SATA over SAS with expanders where I have seen or heard of the problem, I would suggest that either it is so exceedingly unlikely for this to occur on an all-SAS solution that it should be ranked up there with 'the system will just spontaneously turn into a shrubbery' (and suggesting otherwise is, in fact, the definition of 'spreading FUD') OR the bug you're mentioning (but not actually providing links to or the ID of) is not, in fact, the cause of the SATA over SAS issues.
So your position is that even though the Oracle folks have uncovered and fixed the real problem (...)
I've got no real love of Oracle or what they've done to Sun since the purchase - but I would LOVE for them to put Garrett, et al, into a position where they need to eat them words.
So your position is that even though the Oracle folks have uncovered and fixed the real problem, even though it explains the symptoms you attribute to the SAS/SATA interaction (in fact, explains it better than the rather contrived version you are standing upon- including an explanation of why the SAS/SATA interaction triggers the race), and just because you haven't seen it, its not there? Interesting. I sure do hope you approach here is not typical of Nexenta's support team.
Because if is typical then your customers have every right to be frightened out of their mind right now by this arrogant, head-in-the-sand attitude.
@ _Gea what naming does your system that faulted use?
What I am getting at wondering is... Is it in any way limited to only those controllers who under ZFS name the drives as WWN names?
.