SUMMARY: DiskSuite RAID's across machines

From: Daniel Baldoni <dbaldoni_at_iinet.net.au>
Date: Thu Oct 02 2003 - 03:03:35 EDT
G'day folks,

(Here's the second of my delayed summaries <grin>)...

First, here's my original question:
>We have recently built a machine with 2 internal disks (fully mirrored) and
>an external SCSI enclosure (6 disks, 4 in a raid5 - with 1 hot spare and 1
>"blank").  At present, there are metadbs on all 8 drives - with the two
>internals each having 4 copies (so that the box can boot without the
>external enclosure) and the externals 1 each (for a total of 14 replicas).
>
>Now, the client wants a second machine setup to mirror (no pun intended) the
>configuration of their production box.  But, here's the kicker - we have to
>use the same external enclosure.  So, is there some way to tell the second
>machine about a raid5 without having to reinitialise the disks if/when we
>plug them in (losing 50+GB data is not an option <grin>).
>
>Also, I'm interested in people's comments about the setup of the metadbs.
>Is there a better way?  We can't just put replicas on the external enclosure
>(as whichever machine doesn't have it won't boot).  We can't just put them
>on the internal disks in case of failure.

Considering the content of some of the responses I got, I could probably have
been a lot clearer in my original posting.  Let me clarify a few things:
	The two machines, A and B, are not networked together in any way.
	Machine B will be a "hot-swap copy" of A, kept locked away in another
	building so it can be put in place quickly if A should fail.
	The enclosure housing the raid is just a desktop multi-pack, not an
	A1000 or similar.
	We only have DiskSuite, not a full-blown install of Veritas,
	available.

Okay, so here's how to do it...  RTFM.  :-(  Moving a raid from one machine
to another is simply a matter of remembering to use the '-k' switch to
metainit when the raid is defined the second time.  There is a major caveat
with this as DiskSuite will "mark" the raid as OKAY even if it crashed
horribly on the other system.

Now, for the second part of my question - and this is where things get very
tricky.  How to spread out the database replicas to try and ensure maximum
availability.  Here again, the purpose of each machine has to be borne in
mind - so we have currently only gone with the following:
	2 replicas on each of the 2 internal drives.

Here's the reasoning:
	If the enclosure is unavailable anyhow, the machine won't boot
	multi-user as (at least) one fsck will fail.  This is okay as the
	client will want to occasionally power up the backup system to
	do patching, reconfigurations, etc. to keep it synced with the
	production unit.  They may even use a "swap" approach wherein they
	make the changes on the backup and do a quick change-over at a
	convenient time - temporarily reversing the roles of the two
	machines.
	If an internal drive fails, it will need to be replaced anyhow,
	meaning the current fallback machine can fulfill its role in the
	Contingency Plan.
	If both happen, there has been a major catastrophe of some sort. 
	They will go back to backups with other disks, etc.

	We can't put database replicas on any of the external drives to
	assure the 50%+1 rule of availability as they couldn't tell me
	just how they were going to maintain the drives setup.  Remember,
	there's a raid, a hotspare and a "blank" but that may change at any
	point.

And, one important point to remember with database replicas is that their
placement is truly dynamic - in that, so long as you have somewhere to put
them, you can add/delete them at any time.  The above setup may change over
time.

Oh, just to complicate things a little further, the disks in system B are not
the same size as those in A - they're not even the same size as each other.
Does anybody know exactly what information DiskSuite 4.2.1 (these are Solaris
8 boxes) keeps in the database replicas?  Another concern I had about putting
replicas on the external drives was that some sort of sizing information may
have been written as well as state data (which could lead to corruption of 
the mirrors on the internal drives).

If anybody wishes to discuss the pros and cons of replica placements, feel
free to drop me a line off-list.  We could always submit a post-discussion
SUMMARY back to the list.

As always, thanks for your time folks - in particular, kudos to Shridar
Bachu, Hichael Morton, Kevin Buterbaugh, Colin Bigam and Craig Burtenshaw.

Ciao.

-- 
-------------------------------------------------------+---------------------
Daniel Baldoni BAppSc, PGradDipCompSci                 |  Technical Director
require 'std/disclaimer.pl'                            |  LcdS Pty. Ltd.
-------------------------------------------------------+  856B Canning Hwy
Phone/FAX:  +61-8-9364-8171                            |  Applecross
Mobile:     041-888-9794                               |  WA 6153
URL:        http://www.lcds.com.au/                    |  Australia
-------------------------------------------------------+---------------------
"Any time there's something so ridiculous that no rational systems programmer
 would even consider trying it, they send for me."; paraphrased from "King Of
 The Murgos" by David Eddings.  (I'm not good, just crazy)
_______________________________________________
sunmanagers mailing list
sunmanagers@sunmanagers.org
http://www.sunmanagers.org/mailman/listinfo/sunmanagers
Received on Thu Oct 2 03:03:30 2003

This archive was generated by hypermail 2.1.8 : Thu Mar 03 2016 - 06:43:20 EST