Register a SA Forums Account here!
JOINING THE SA FORUMS WILL REMOVE THIS BIG AD, THE ANNOYING UNDERLINED ADS, AND STUPID INTERSTITIAL ADS!!!

You can: log in, read the tech support FAQ, or request your lost password. This dumb message (and those ads) will appear on every screen until you register! Get rid of this crap by registering your own SA Forums Account and joining roughly 150,000 Goons, for the one-time price of $9.95! We charge money because it costs us money per month for bills, and since we don't believe in showing ads to our users, we try to make the money back through forum registrations.
 
  • Post
  • Reply
rage-saq
Mar 21, 2001

Thats so ninja...

unknown posted:

How many people are doing boot-from-SAN (ie: no hard drives on the physical server)?

I'm absolutely loving it, you can't beat the fact of screwing up a boot disk, and to fix you simply present + remount it as a regular volume on a different computer to easily fix. All from a 1000 miles away.

I've done a few good sized network upgrades. At two remote co-located sites we did a blade chassis and a small SAN and did boot-from-SAN. Each site has an extra blade on hand in case a server were to fail so we could remap the LUN and get it back up and running remotely in a short period of time.
We did that with the customer before they had gone through testing and validation of virtualization instead, which is how all new deployments are done.

Adbot
ADBOT LOVES YOU

rage-saq
Mar 21, 2001

Thats so ninja...

KS posted:

Update:

We shut down our entire SAN last weekend and brought up one linux server, then one 2k3 server. Performance was identical to the performance we get in the middle of the day at peak load.

We found some benchmarks here .These are reads in MB/sec over numbers of I/O streams.



This jibes with what we're seeing. How is that single stream performance anywhere near acceptable? I can throw 4 SATA disks in a software RAID-5 and beat that read performance.

What are the strategies, if any, we should be implementing here? Striping volumes across multiple vdisks? Tweaks to increase the number of i/o "streams" per server? How will we ever get acceptable ESX performance?

:smith:

Open up a case with HP, something is seriously not right here.

rage-saq
Mar 21, 2001

Thats so ninja...

unknown posted:

What are people using for doing their I/O tests of boxes?

bonnie++ or just straight 'dd' or any other package?

I prefer IOmeter. Here are some ideas towards creating some workloads to help evaluate storage performance.

code:
Shamefully ripped from the vmware forums

######## TEST NAME: Max Throughput-100%Read
size,% of size,% reads,% random,delay,burst,align,reply
32768,100,100,0,0,1,0,0

######## TEST NAME: RealLife-60%Rand-65%Read
size,% of size,% reads,% random,delay,burst,align,reply
8192,100,65,60,0,1,0,0

######## TEST NAME: Max Throughput-50%Read
size,% of size,% reads,% random,delay,burst,align,reply
32768,100,50,0,0,1,0,0

######## TEST NAME: Random-8k-70%Read
size,% of size,% reads,% random,delay,burst,align,reply
8192,100,70,100,0,1,0,0

rage-saq
Mar 21, 2001

Thats so ninja...

KS posted:

performance woes

Actually I just remembered something, I had some similar performance issues on some BL25ps because of some stock settings of the HBA driver/BIOS. Are you using Qlogic or Emulex?
What is your maximum queue depth and/or execution throttle? You might want to try messing with that figure to see if you can improve your single server scenario.

edit: Also, theres a bunch of storage guys in #shsc on irc.synirc.org. You should stop by and chat we might be able to give you other pointers.

rage-saq fucked around with this message at 20:06 on Sep 2, 2008

rage-saq
Mar 21, 2001

Thats so ninja...

echo465 posted:

Anyone using a Promise SAN? We're using a Promise VTrak M610i, which is a 16-SATA, iSCSI box for around $5k. Populate it with 1TB hard drives (that are really only 930MB or so), and you've got 13TB of RAID-6 storage for around $7500.

Get ready to have fun when the controller shits itself and annihilates all the data on your array, and Promise has no idea why and offers no recourse for fixing it or preventing it from happening again (which it will).

I've had 3 different customers who had differing Promise units and they all did something along these lines happen, and Promise basically told them to go gently caress themselves.

I can't tell people enough to avoid this kind of crap.

rage-saq
Mar 21, 2001

Thats so ninja...

Alowishus posted:

code:
LUN Capacity(Megabytes):    102400
LUN Capacity(Blocks):       209715200
... I can infer that the blocksize is 2048, but is that something I can rely on to always be the case? If not, is there an easier way to get it out of Navisphere?

I'm not an EMC guy (I've had very little experience with it actually) but I'd definitely say you are at a 2k block size. Lots of larger SANs don't give you the option of your block size. HP EVA, Netapps and I think HDS are all 4k blocks and you can't change it.

rage-saq
Mar 21, 2001

Thats so ninja...

H110Hawk posted:

I come seeking a hardware recommendation. It's not quite enterprise, but it's not quite home use, either. We need a bulk storage solution on the cheap. Performance is not a real concern of ours, it just needs to work.

Our requirements:

Hardware raid, dual parity preferred (RAID6), BBU
Cheap!
Runs or attachable to Debian Etch, 2.6 kernel.
Power-dense per gb.
Cheap!

To give you an idea, we currently have a 24-bay server with a 3ware 9690SA hooked up to a SAS/SATA backplane, and have stuffed it full of 1TB disks. People are using SFTP/FTP and soon RSYNC to upload data, but once the disks are full they are likely to simply stay that way with minimal I/O.

We are looking for the cheapest/densest cost/gig. If that is buying an external array to attach to this system via SAS/FC, OK! If it's just buying more of these systems, then so be it. I've started poking around at Dot Hill, LSI, and just a JBOD looking system, but I figured I would ask here as well, since LSI makes it hard to contact a sales rep, and Dot Hill stuff on CDW has no pricing.

The ability to buy the unit without any disks is a big plus, we are frequently able to purchase disks well below retail.

I need about 50-100tb usable in the coming month or two, then it will scale back. I am putting these in 60amp/110v four post 19" racks.

Edit: Oh and I will murder anyone who suggests coraid, and not just because it is neither power dense nor hardware raid.

Get an HP SmartArray P800 card and then attack up to 8 MSA60 12xLFF enclosures.
P800 is about $1k, each shelf is about $3k and then add your 1TB LFF SATA drives and you are good to go. If you need more attach another P800 and more shelves etc.

rage-saq
Mar 21, 2001

Thats so ninja...

Alowishus posted:

Just out of curiosity, in an HP environment, how does one add their own 1TB LFF SATA drives? Is there a source for just the drive carriers? Because using HP's 1TB SATA drives @~$700/ea makes this quite an expensive solution.

I don't know if theres a SKU for the correct drive trays, generally I just tell customers to buy the HP drives and thats the end of it.

rage-saq
Mar 21, 2001

Thats so ninja...

Catch 22 posted:

I didn't calculate that number, that comes from my "Storage Specialist" at Dell telling me my peak IOSP to use a Equallogic. I calculated MY IOPS for my environment, and I added on to my real number to account for the RAID overhead.

I assure you I wont argue if someone calls "bullshit". I don't trust anyone trying to sell me poo poo, thus why I am here.

Its a nice device, but 60K for a SATA SAN just boggles my mind.
He did show me a 2000 mailbox Exchange 2007 environment running off a SATA one though.

Yeah, if he's telling you 3k IOPS with that equalogic, he clearly has NO idea what he's talking about. Thats just peak IOPS for that unit under the right configuration, which is NOT how someone who knows how to design storage is going to talk about the speed of an array with you.
Someone that knows what they are doing and isn't just some retarded sales drone parroting marketing information and talking points given to him in his 'training classes' is going to with you about what kind of IOPS you currently push, where you are going, and what kind of setup you need to achieve that.

That being said I've setup a few Equalogics for customers who purchased them before engaging my services, and I don't really like them. Silly setup constraints like not being able to use all of the drives in the chassis in one array and other things. Their cabling setup of only 2x 1GbE despite having 3 ports was odd, their web management page was alright but left a lot to be desired.
One of the customers who I setup said Equalogic 2 months is already going to be moving to a better SAN because it just wasn't cutting it as configured.

And about his comment of an Equalogic box running a 2000 user Exchange 2007 environment, I'm going to either call bullshit or say it ran unacceptably slow. I've done a few 2000 user Exchange designs, and the last design will use 16x 146gb 15k HDD in RAID10 for the database and then a small LUN on a separate 6x300gb 15k RAID10 disk group for the log files. I really doubt a single equalogic box would be able to handle it acceptably unless it was a REALLY lightly used Exchange server.

rage-saq fucked around with this message at 18:17 on Sep 19, 2008

rage-saq
Mar 21, 2001

Thats so ninja...

brent78 posted:

Anyone else going to be at the HP Storage Symposium in Colorado Springs next month? Send me a PM, I'm staying at the Broadmoor (as is everyone else I imagine)

Hmm, are you going through a distributor or through HP? One of my distributor guys whispered something about another storage event soon (last one was in SF).

rage-saq
Mar 21, 2001

Thats so ninja...

Kullrock posted:

What is LFF short for?, Can you use regular 1TB sata disks on those, or do you need to pay out your rear end for HP's own disks?

LFF = Large Form Factor aka 3.5"
SFF = Small Form Factor aka 2.5"

I believe you could use any old 1TB drives in there as long as you found the empty trays and purchased them, but I don't recommend it to my customers as they wouldn't get HP warranty coverage.
Plus HPs has Dual Port 1TB 7200 SAS and have a more uniform higher performance and reliability than standard SATA. The Dual Port means your shelves can have two data paths all the way from the drive to shelf to controller(s) to give you fiber-channel like availability for a fraction of the cost.

HP's LFF drive lineup looks like this right now.
code:
SP = Single Port
DP = Dual Port

Online SAS
HP 72GB  3G 15K SAS SP 
HP 146GB 3G 15K SAS SP 
HP 300GB 3G 15K SAS SP
 
HP 72GB  3G 15K SAS DP 
HP 146GB 3G 15K SAS DP 
HP 300GB 3G 15K SAS DP 
HP 450GB 3G 15K SAS DP 

Midline SAS
HP 750 GB 3G 7.2K SAS DP 
HP 1TB    3G 7.2K SAS DP 

Midline SATA
HP 1TB    3G 7.2K SATA SP 
HP 750 GB 3G 7.2K SATA SP 
HP 500 GB 3G 7.2K SATA SP 
HP 250 GB 3G 7.2K SATA SP 

rage-saq
Mar 21, 2001

Thats so ninja...

Mierdaan posted:

I was originally talking about the Dell MD3000i; I don't know what EQL product number that is (it is EQL hardware, right?

No, the MD3000i is a box from LSI. Equallogic are totally different products, but I wouldn't buy either. They are both rather immature products in a large number of ways (web management, strange limitations and non-competitive price/performance).

rage-saq
Mar 21, 2001

Thats so ninja...
Catch-22 what kind of configuration are you looking at and what kind of features? I can estimate to you how much a good HP setup would cost, and worst case you could use that to get a better price from EMC.

rage-saq
Mar 21, 2001

Thats so ninja...

Mierdaan posted:

Thanks rage-saq. What would you say is the best bang for your buck in the $15-$25k range? As previously mentioned we don't currently need anything with crazy good performance, but we don't want to handicap ourselves later. Only need about 3-4TB usable space, with the ability to scale up later.

What kind of features? How many hosts? What kind of disk configuration?
HP has a really great lineup with the MSA2000. There is an FC model, iSCSI model and SAS (still a SAN, think FC without FC) model with both single and dual controller/fabric versions and they are all pretty awesome from a management and price/performance standpoint.

rage-saq
Mar 21, 2001

Thats so ninja...

Mierdaan posted:

Snapshots are really all we'd be looking for right now. Don't need replication as I'm still working on getting the first device in the door :(

20 or so hosts would be reasonable for the forseeable future.

Disks, we could realistically go for a mix of large SATA and smaller faster SAS drives, but our first stab at carving up 15 spindles in the MD3000i was:
1) RAID 10, 4x 300GB 15K RPM SAS drives: SQL DB
2) RAID 10, 4x 146GB 15K RPM SAS drives: SQL transaction logs
3) RAID 1, 2x 73GB 15K RPM SAS drives: Exchange transaction logs
4) RAID 5, 4x 450GB 15K RPM SAS drives: VMimages, light file server use
(with 1 spare drive for whatever RAID set we decide most needs it)

The MSA2000 series can do snapshots and volume copies, but not array based replication. Additionally, the MSA2000 series is a block-level virtualized storage system, so you don't necessarily have to carve up disk groups based off I/O patterns like you do with traditional disk-stripe arrays.

MSA2000 main controller shelf is 12 disks, and additional shelves are also 12 disks.

Also, transaction logs are nearly 100% sequential, and an array that operates in nearly 100% sequential mode will perform ~2.5x as fast, so you can really get by with a lot fewer disks and consolidate your LUNs there.
You could probably get the space and the performance you need by going with 8x 450gb in RAID10 and 4x146GB (logs) in RAID10.

At 20 hosts you are looking at either FC or iSCSI, and given your tight budget you'll probably want to go iSCSI.

For reference, a dual-controller MSA2012i with 8x450gb and 4x146gb would come in around 17k or so.

rage-saq
Mar 21, 2001

Thats so ninja...

Catch 22 posted:

Opps, sorry forgot that.

12 x EMC 146GB 15K SAS DRIVE AX-SS15-146
12 x EMC 750GB 7200RPM SATA II DRIVE

Whats your AIM? Some of those other part numbers you mentioned don't really mean anything to me because I'm not well versed in EMC part numbers.
Or you can come into #shsc on irc.synirc.org, we've a few storage pros in there quite regularly.

rage-saq
Mar 21, 2001

Thats so ninja...

BonoMan posted:

I guess I can ask this part of the question as well...what's a good pipeline for backing up from a SAN or NAS?

They want server A to be current work. Server B to be archived work. All of that work backed up to tape for offsite storage and then also have physical backups in house (DVDs or whatever). They would also like a synced offsite server somewhere for fast restoration only.....they really won't pay for jack. :( Or at least not the massive costs it would cost for that.

You need to hire a consultant to come in and help you understand what you want. You could be talking about something in the 40-50k range (unless you need new servers too) or you could be talking about something in the 300k+ range. A qualified consultant will help you figure out what you want vs what you want to pay for.

rage-saq
Mar 21, 2001

Thats so ninja...

BonoMan posted:

Yeah we have one coming in next week. Like I said this is just me putting feelers out there to see what other folks use.

It depends on a lot of aspects. A few things to start thinking about now.

1: Recovery Point Objective, what is acceptable data loss, the closer to 0 you get, the more money you spend. 0 data loss is almost cost-exorbitant, and requires a few hundred grand in equipment to achieve and its still not guaranteed.

2: Recovery Time Objective, what is an acceptable downtime. Again, the closer to 0 you get the costs go up exponentially. An RTO of 0 means server redundancy, storage redundancy (two SANs) and application redundancy (example: Microsoft Exchange Continuous Cluster Repication).

3: IOPS. The amount of IOPS you need is going to ultimately determine the expense of the array. If you've got something that needs a sustained random IOPS level of about 2000 IOPS on a database set of 300gb or more, you are talking about getting about 20 disks just for that one application.

rage-saq
Mar 21, 2001

Thats so ninja...

Catch 22 posted:

Call EMC, they will come out, do metrics for you, go over everything they think you need, and they will do you right. You can then use this info to compare at other products. Be frank with them and do not think of them as "demoing" a product to you.

You can Even do this with Dell, and they will shove a Equallogic down your throat but they can give info and help you figure out what you need.

This is all free. Just call some vendors.

EMC actually came out onsite and did some performance monitoring to determine IOPS usage patterns before you gave them any money?

I've not actually heard of them doing this, just coming up with random guesstimates for customers based off a little input from the customer. They were horribly wrong (short by about 40% in some cases) that I ended up fixing their order at the last minute before they placed their order (but sometimes after and having to fix it by buying more disks).

Moral of the story: You can't cheat by guessing at IOPS patterns, you really need to know what your usage patterns look like.
Some applications (like Exchange) have decent guidelines, but they are just that, guidelines. I've seen the 'high' Microsoft Exchange estimates be short by about 50% of actual usage, and I've also seen peoples mail systems be 20% under the 'low' guideline.
SQL is impossible to guideline, you need to do a heavy usage case scenario where you record lots of logs to determine what to expect.

rage-saq
Mar 21, 2001

Thats so ninja...

Mierdaan posted:

Can you explain this a little more? I've always felt like I was missing something by carving up disk groups like I did before, so I'm glad to know I am. I just don't understand quite what :)

Reposting this question on a new page :) Is rage-saq just talking about the ability to automatically move more frequently-used data to faster disks? That makes sense if you have slower spindles in your device, but if you cram the whole thing full of 15k RPM SAS drives I'm not quite sure what this accomplishes. Again, totally missing something :(
[/quote]

Well, there are two basic RAID types used today.
The old way: Disk raid. A 32kb or 64kb or 128kb stripe (sometimes per lun, sometimes not) runs through the entire disk group. Sometimes you make multiple disk groups and group those guys together for performance. Different vendors have different strategies.
What are the downsides? The stripe must be maintained, so performance and reliability follow traditional models and consolidating I/O patterns to particular disk groups (random vs sequential) is of great concern.

The newer way: Block raid. Data is virtualized down to blocks, which are then spread out over the disk group. The advantage is the blocks aren't tied to particular disks, so they can be spread out to maintain redundancy and improve performance as needed. That means you can mix I/O patterns without seeing a performance penalty because the blocks will get spread out more and optimized so more disks are driving the I/O.
It also means it doesn't matter so much that all the disks match the same disk size and spindle speed, but to make your RAID meet sizing and redundancy requirements you won't be able to use the fullest extent of your capacity when you mix spindles.
Not all block storage systems will automatically migrate data over different disk spindle speeds like Compellent and 3Par, that is still fairly new. I know HP is working on including migration into EVAs eventually, but right now its still a best practice to have the same spindle speed per disk group in an EVA. Drobo uses virtualized storage to spread out redundancy blocks everywhere so they can get their crazy recovery thing going.
EMC unfortunately is very pompous and is under the misguided opinion that block virtualization is a bad which is why they don't have it. A lot of industry experts disagree.

rage-saq
Mar 21, 2001

Thats so ninja...

ghostinmyshell posted:

I'm not happy with Lefthands support. As soon as LH can identify the problem is not in relation to their software, you enter the bermuda triangle dealing with HP for the hardware side of things.

We had their 4 hr premium support, so LH calls up HP to dick around with them to get a tech sent out. I mean seriously, I was on the line listening to the HP rep give the same poo poo to our LH case manager that you or any other person calling for tech support would get.

Eventually, HP send us a technician the next day with the wrong parts and who knows absolutely nothing about the problem we are having with the software or the hardware. gently caress I don't care by this point who fixes the problem.

LH then plays the blame game with HP and it's going back and forth. I really wanted to tell our LH case manager, "No, listen you fucks we pay you for the support, therefore the burden is on you guys to fix our problems, quit loving around here like a smug rear end in a top hat."

After some calls to higher ups, we finally got the right hardware, a decent tech and the problem resolved... only two days later.

Sounds like someone is wishing they got a 6 Hour Call To Repair warranty instead :)
Its mighty expensive though, nearly double 4 hour response!

rage-saq
Mar 21, 2001

Thats so ninja...

oblomov posted:

Rage, can you clarify on this? Is this through HP or Lefthand support? Personally, I think I am going to go with NetApp after all.

Well in a few weeks their support will be through HP.
6 hour Call To Repair is expensive because it is the highest level of support you can get. You call in an issue and HP basically promises it is resolved in 6 hours. That means they stock spare parts for you at the local warehouse, and make sure they have enough certified techs in your area to cover you should a problem arise. This is basically pulling out all the stops so your poo poo gets fixed ASAP.

rage-saq
Mar 21, 2001

Thats so ninja...

oblomov posted:

Hmm... I will hit up my rep on this. The other route is frankly unacceptable. One of big reasons even at looking at something other then NetApp is less downtime and quicker potential fixes. If it takes 2 days to repair a SAN, that's simply unacceptable. The fact that their support is going to be through HP is what concerns me most actually. IMO, HP support is simply not up to par considering their pricing.

HP support options like 4 hour 24/7 etc is just as good as anyone elses these days on Standard equipment, if you really need to have insane support you buy 6 hour CTR. LeftHand should be getting shelved under Standard equipment so the same Proliant level warranty options will apply.

If you REALLY want crazy support, you need to step it up to Enterprise equipment, stuff like EVA's, ESL's and XP's have insane warranty options. This is the kind of equipment that competes directly with Netapp, as Netapp predominantly makes Enterprise level equipment.
EVA's are awesome (installing one right now as a matter of fact) and have all sorts of warranty/support/prefailure replacement + auto dialhome that just don't appear on the Lefthand/MSA level Standard storage lines.

If you get an XP you can get a support contract with HP for 7 9's uptime. Thats like 6 seconds of downtime per year. This is what runs the NASDAQ and thats the kind of warranty they have.

rage-saq fucked around with this message at 02:39 on Jan 5, 2009

rage-saq
Mar 21, 2001

Thats so ninja...

Misogynist posted:

IT people who have no balls (or female equivalent).

They are called Thatchers. :colbert:

skullone posted:

And whatever NAS/SAN you get, it'll suck. There will always be odd performance problems that you'll have to spend hours troubleshooting before your vendor will listen to what your saying, only to have them say its a known problem, and a patch will be ready in a few weeks.

I'm not bitter or anything

The fact that there are consultants/vendors out there that allow this kind of behavior is appalling. You are dropping a lot of cash for a very advanced piece of equipment that is just supposed to work, if it doesn't you should return it and get something that does.
I do enterprise storage design/consulting/implementation primarily around HP products and I can honestly say all of my deployments work 100% as advertised with no mysterious performance/reliability problems. Thats the whole loving point of doing this. If the product couldn't deliver as advertised I would be the first one trying to get the customer a refund as well as not recommending it in the future.

My personal opinion is that this is what you get when you go with generic server equipment and then use some kind of general purpose operating system + software package to accomplish this kind of low level stuff. A lot of Sun's entry products utilizing ZFS seem to fit this kind of bill along with other stuff I'm not a fan of like LeftHand etc. Your mileage may vary of course.

rage-saq
Mar 21, 2001

Thats so ninja...

Mierdaan posted:

Well, we've been hosed.
<stuff>

Reason number 5234634 to not deal with those big fullfilment warehouses like CDW, PC-Mall etc for configs.
The guy you are talking to really has no clue about any high level technical stuff, he's just a guy that knows sales stuff and is 'good with computers'. Always always always hire a consultant to come out and do the design/config work for you, its not free but at least it will be right and you will have significant recourse if they screw up the config.
Because you took your PC-Mall config which was hosed up, directly to another vendor who didn't have any part in the config, you are pretty much hosed.

rage-saq
Mar 21, 2001

Thats so ninja...

mkosmo posted:

Hey Storage Gurus, I have a question for you if you will permit:

At work we have an EMC Clariion (CX3-80) now replacing our old Intransa and a couple Isilon units (which have performed very poorly for our I/O needs).

EMC promised us >=350 megabytes per second per datamover and we're not really seeing that despite having been working with engineers for months. Also it appears there is no jumbo framing on the 10GbE interfaces which could cause us some poorer performance. In addition, getting CIFS and NFS working on one file system cooperatively proved to be a hassle.

Any idea whats up with that? What other issues have you seen with EMC and performance or otherwise?

First my two cents, your EMC engineer who sold and implemented the system for you should have properly designed your system to meet your necessary workload performance requirements.

Now on to the real question of why this is a difficult thing to achieve.

>350MB/s is quite a bit of throughput, but throughput alone is not the most important factor when it comes to disk performance. It can actually be extremely misleading. Disk performance is a very careful balancing act of the right raid level at the right block size for an optimal IOPS & throughput level to best meet your I/O pattern.

For example

350MB/s @ 8 IOPS would be a pretty poorly performing disk system, whereas 80MB/s @ 5000 IOPS could be an extremely well performing disk system.

A few things that will affect your performance that you'll want to look into, be sure to use IOMeter to do your performance testing, anything else is basically full of lies (I'm looking at you HDtach)

1: I/O pattern. This means what % is read, what % is write, and what % is random vs sequential.
2: Block sizes. The block size of your stripe (I think EMC calls this element size?) and the block size of your filesystems partition.

Per the same number of disks in a disk group you will have to balance the block size for both IOPS and throughput (MB/s).
The smaller the block size the more IOPS you are going to get, but at a lower MB/s. Conversely the larger the block size the more MB/s you are going to get but at a much lower IOPS.

Additionally you are going to get better IOPS AND MB/s the more sequential your workload is, and also even more if your workload is more read than write.

8kb block size on both the disk and filesystem size is one of the better balanced configurations. With a 24x 450gb 15k disk group on an HP EVA4400 I've gotten in excess of 6000 IOPS (some of the figure would be cached performance) at 250-300 MB/s in certain access patterns. Database servers and anything that has a highly random and mildly write heavy, like operating system drives and things like that.

Fileservers you can increase your block size to 32kb or higher and get good throughput but at a lower IOPS rate (which is generally fine for a file server).

IOMeter is going to become your best friend for evaluating your disk/filesystem configurations to see if you get the performance you need to meet your workload.

rage-saq fucked around with this message at 22:30 on Feb 4, 2009

rage-saq
Mar 21, 2001

Thats so ninja...

InferiorWang posted:

Thanks fellas. What I'd really like to do is have an iSCSI SAN, 2-4 TB. I'd like to host a modest amount of vmware guests on it hosting primarily file shares, home directories, and groupwise email for the staff numbering 250-300. Then I'd like to do it all over again with a completely redundant SAN at one of our other schools in some sort of fail over configuration. Being able to do snapshots and have some sort of monitoring dashboard is something I'd like as well.

My problem is I can't even get anyone to listen to me about doing this and getting away from bare metal machines for our critical data without having some semblance of a dollar figure attached to it. I have to approach this rear end backwards from how most normal people would approach it.

Pick a big number and get them to approve that, then show off how good you are by coming in under budget!

rage-saq
Mar 21, 2001

Thats so ninja...

InferiorWang posted:

I might be better off saying it's cheap and ending up over budget. Otherwise I'm likely to get this response: :frogout:

I'll look into the CDWG route. I do have a rep and he's generally pretty good.

I'd advise against getting any kind of advice on important system design from what is essentially the best buy of product fulfillment centers. Call in a consultant and engage with them in a discovery and project assessment for virtualizing your infrastructure.

rage-saq
Mar 21, 2001

Thats so ninja...

BonoMan posted:

Being new to SAN type storage...what are options that don't require per license usage?

One of our companies has a FC SAN that has Stornext or whatever and it's like 3500 per computer. Do Dell EqualLogic's that are iSCSI require licenses per user?

The use of StorNext is totally not required for any SAN usage at all, it is for very specialized use scenarios.
When you map a volume off a SAN to a system, it appears as a any regular local drive it has exclusive block level access to. SANs abilities to share those same volumes between two servers is great, but without a shared filesystem it would be useless. Two systems would write to the file table at the same time and just destroy the whole thing.

StorNext provides that shared filesystems so all the computers that have that volume mapped locally can use a centralized locking system so they don't write to the FAT/same files at the same time where you would get a destroyed filesystem.

rage-saq
Mar 21, 2001

Thats so ninja...

bmoyles posted:

Say you need 100TB of storage, nothing super fast or expensive, but easily expandable and preferably managed as a single unit. Think archival-type storage, with content added frequently, but retrieved much less frequently, especially as time goes on. What do you go for these days?

If you want file-presentation instead of block-presentation from your unit, the new HP Extreme Data Storage 9100 is pretty badass. Really low cost per GB, very fast NFS, CIFS, managed through one console. Its also extremely dense due to their super fancy new disk shelves that hold 82 LFF disks in a 5u shelf. Dreamworks just purchased a few petabytes of it.

rage-saq
Mar 21, 2001

Thats so ninja...

Erwin posted:

Anybody familiar with Scale Computing? They caught our eye because they were originally in the financial market, which is what we do (even though they failed at it). Their system starts with 3 1U units, either 1TB, 2TB, or 4TB per unit, and then you can expand 1U at a time, mixing and matching pur-U capacities.

Skip it. IMO the only clustered (as opposed to traditional aggregate arrays like HP EVA, EMC CX, netapp etc) systems worth getting into are LeftHand or some kind of Solaris ZFS system.

rage-saq
Mar 21, 2001

Thats so ninja...

1000101 posted:

EMC Recoverpoint, and potentially NetApp Snapmirror can both give you application consistent replication.

Depending on distance between endpoints, RPO, and latency you may need to live with crash consistent replication for business continuity/disaster recovery.

How does oracle's built in replication handle it? Does it wait for the other side to acknowledge a write before committing it to logs at the local site? How would say 70ms of latency impact the performance of the production database?

I'm pretty sure synchronous replication (where it ACKs a mirrored write before continuing on the source side) and 70ms of latency would destroy nearly any production database performance. At a certain point you can't overcome latency.

rage-saq
Mar 21, 2001

Thats so ninja...

three posted:

Hrms. What would you suggest as the best route to copy the database off-site to another SAN? We actually just want a snapshot every hour or so, not live replication. I'm very new to Oracle, but have good MySQL experience, and to a lesser extent MSSQL experience. Could we just do similar to a dump and import like would be done with MySQL? Although that wouldn't be optimal since we'd have to lock the entire DB to dump it, no?

They bought these equallogic units before I started here because the Dell rep convinced them syncing Oracle would be as simple as replicating using the equallogic interface. :sigh:

This kind of advanced high availability is not as simple as many sales people would like it to sound. Some options like application specific mirroring (I'm not an Oracle pro, but I know MS-SQL 2005/2008 have both sync and async database mirroring with failover) would allow you some technical avenues towards improving the situation.

If the people who bought it weren't smart enough to know what they were asking for, or didn't know enough to ask the right questions and not have the wool pulled over their head, then theres not much you can do from a business/customer service standpoint. However, if the RFP/requirements sheet/whatever was specific enough it called for these specific features, and you didn't get them due to whatever kind of fuckup then you should take some hints from comments people have already made and make DELL fix this.
As people have already mentioned, find out WHO made the promises, get emails of them if possible, and hold them to it. If you are a vendor/consultant/etc and you make claims about XYZ with vanilla icecream on the side, and it doesn't do any of that, you can be sure as hell you can LEGALLY get out of paying agreed upon price. Thats how this industry works.

If you have already paid it they would be REALLY REALLY DUMB for them to not do EVERYTHING they can possibly do to make it better. These kinds of empty promises and bad service followup is every competitors and lawyers dream situation and every product/service/sales managers worst nightmare.

I'm a consultant and you better bet this is one of the highest things on my list of "Things Not To Do". I do everything I can to make sure I understand the technical requirements, and make sure the customer understands the differences and why its important these things are defined and not glossed over. If I promise X solution and can only deliver Y because I hosed up for whatever reason, you better bet my boss is going to chew my rear end out because he has to pay to fix it and I may have to be working on a new resume sooner than I had wanted.

I will say I'm not surprised, DELL internal equallogic sales reps are notoriously clueless (getting someone barely qualified for sales, and barely qualified for basic end user technical support doing advanced enterprise design is their SOP) and I've had many customers buy the heaping load of bullshit they get fed by somebody who clearly doesn't understand what they are talking about only to get disappointed later.
Great specific examples of this are when they get sold a single SATA shelf and are promised 50k IOPS because that is what the maximum value on the specs sheet says. Meanwhile the sales rep says its going to meet the performance requirements for a write heavy SQL database for this and god knows what other reason. If you don't press to get real technical knowledge they don't throw someone qualified at you.

1000101 posted:

Clearly this is the case but I'm not sure how Oracle handles it. Is Oracle going to acknowledge every write? If so then yeah 70ms would make the DBA want to blow its brains out. Or does it just ship logs on some kind of interval? More of an Oracle question than a storage question but it would be helpful in deciding if you should use array based replication or application based.

I suppose thats the real question of *WHERE* you are doing your replication and what your application supports. If Oracle doesn't care how it got the database so long as it was crash consistent with the log files you could get away with some pretty dirty array based replication techniques that do a synchronous replication that only amounts to *COPY BLOCKX, WAIT FOR CONFIRMATION ON BLOCK X BEING WRITTEN BEFORE SENDING BLOCK Y* as opposed to some more advanced things like array agents for oracle or oracle itself doing the replication.

Different techniques allow for different options but at a fundamental level you can't have *true* synchronous replication that doesn't hold up processing on the source end until the target matches the source. Otherwise its not no-data-lost-crash-consistent which then defeats the point of synchronous replication and that kind of scenario where something like 70ms delay would really kill you. Its just the nature of the beast.

rage-saq
Mar 21, 2001

Thats so ninja...

Insane Clown Pussy posted:

So, calling around trying to find a replacement for our old Lefthand boxes. I need something in the 2-4 TB range for 2-3 ESX hosts running a typical 6-8 server Windows environment (files, exchange, sql, sharepoint etc)

Just get another LeftHand node, that way you can use your existing LeftHand stuff for secondary storage / backups / etc. Its a good platform and if your infrastructure is right it will perform more than adequately for the kind of environment you are talking about.

Adbot
ADBOT LOVES YOU

rage-saq
Mar 21, 2001

Thats so ninja...

EoRaptor posted:

Most have these have already been answered, but I'll add some things Dell/Equallogic isn't very public about :

Equallogic 'groups' are actually grids, not clouds. This might seem like a small difference, but if a member of an equallogic group goes offline, all storage that crosses that member will also go offline. By default, that's all storage in the group, so a failure of a single device can take down much more than you'd think. Dell puts forth that units are five 9's devices, but human error could still have large effect.

Seriously? So as you build your EQL "grid" up you are introducing more single points of failure? Holy cow...

  • 1
  • 2
  • 3
  • 4
  • 5
  • Post
  • Reply