[Ocfs2-users] Cluster setup
Pavel Georgiev
pavel at netclime.com
Thu Oct 11 02:47:36 PDT 2007
On Thursday 11 October 2007 00:24:41 Alexei_Roudnev wrote:
> Yes, it can be done.
>
> Question is in reliability:
> - OCFSv2 is not very stable when it is about millions of files;
> - OCFSv2 cluster tend to self-fence after a small SAN storage glitches (it
> is by design so you can't heal it even if
> you fix all timeouts - just to improve);
> - OCFSv2 + LVM +_ NFS is not well tested territory.
>
> It should work - in theory, IT works practically, under average load and FS
> size. No one knows how it behaves on a very big storage and a very big file
> systems in 1 - 2 years of active usage. I manage ti get stable OCFSv2
> system here, after applying few pathces and discovering few issues, BUT
> I use it on lightly-loaded file system (which is not critical at all)to get
> more statistics on behavior before I wil use it for anything else.
>
> If comparing with heartbeat + LVM + reiserfs + NFS:
This comes a little bit offtopic for an ocfs2 list but, how do I meet my
requirements with just reiserfs, LVM and heartbeat? Do you mean running DRBD
and reiserfs on top of it? This looks like a combination I could use -
currently I have a single storage server which many clients mount, if I
implement that (RDBD + heartbeat + nsf, no lvm) it would meat my requirements
(except for the expanding of storage maybe) and I will not have to make any
changes to the clients which is great. Do you say there is a way to do this
(replicating filesystem between two nodes) with just heartbeat + lvm?
> - all technologies in stack are well tested and heavily used;
> - heartbeat have external fencing (stonith) so it is extremely reliable in
> the long term - it can recover from almost any failure (sometimes it dont
> feel failure, it's true);
> - ReiserFS (or ext3) proved to be very stable on a huge file systems (it is
> widely used, so we dont expect any problems here).
> One problem comes from Novell - since they stoped using it as a default, I
> can';t trust to ReiserFS on SLES10 (because it is not default) but we stil
> can trust into it on SLES9 etc... (where it is default).
>
> Common rule - if you want reliable system, use defaults where possible.
> OCFSv2 + NFS is not default yet (through OCFSv2 improved dramatically
> during last 2 years).
>
> ----- Original Message -----
> From: "Pavel Georgiev" <pavel at netclime.com>
> To: <ocfs2-users at oss.oracle.com>
> Sent: Wednesday, October 10, 2007 1:25 AM
> Subject: Re: [Ocfs2-users] Cluster setup
>
> > How about using just OCFSv2 as I described in my first mail - two servers
> > export their storage, the rest of the servers mount it and a failure of
> > any
> > of the two storage servers remains transparent to the clients. Can this
> > be done with OCFSv2?
> >
> > On Tuesday 09 October 2007 21:46:15 Alexei_Roudnev wrote:
> >> You better use
> >>
> >> LVM + heartbeat + NFS + cold failover cluster.
> >>
> >> It works 100% stable and is 100% safe from the bugs (and it allows
> >> online resizing, if your HBA or iSCSI can add lun's on the fly).
> >>
> >> Combining NFS + LVM + OCFSv2 can cause many unpredictable problems, esp.
> >> on
> >> the unusual (for OCFSv2) system (such as Ubuntu).
> >>
> >> ----- Original Message -----
> >> From: "Brian Anderson" <banderson at athenahealth.com>
> >> To: <ocfs2-users at oss.oracle.com>
> >> Sent: Tuesday, October 09, 2007 11:35 AM
> >> Subject: RE: [Ocfs2-users] Cluster setup
> >>
> >>
> >>
> >> Not exactly. I'm in a similar boat right now. I have 3 NFS servers all
> >> mounting an OCFS2 volume. Each NFS server has its own IP, and the
> >> clients load balance manually... some mount fs1, others fs2, and the
> >> rest fs3. In an ideal world, I'd have the NFS cluster presenting a
> >> single IP, and failing over / load balancing some other way.
> >>
> >> I'm looking at NFS v4 as one potential avenue (no single IP, but it does
> >> let you fail over from 1 server to the next in line), and commercial
> >> products such as IBRIX.
> >>
> >>
> >>
> >>
> >> Brian
> >>
> >> > -----Original Message-----
> >> > From: ocfs2-users-bounces at oss.oracle.com
> >> > [mailto:ocfs2-users-bounces at oss.oracle.com] On Behalf Of Sunil Mushran
> >> > Sent: Tuesday, October 09, 2007 2:27 PM
> >> > To: Luis Freitas
> >> > Cc: ocfs2-users at oss.oracle.com
> >> > Subject: Re: [Ocfs2-users] Cluster setup
> >> >
> >> > Unsure what you mean. If the two servers mount the same
> >> > ocfs2 volume and export them via nfs, isn't that clustered nfs?
> >> >
> >> > Luis Freitas wrote:
> >> > > Is there any cluster NFS solution out there? (Two NFS
> >> >
> >> > servers sharing
> >> >
> >> > > the same filesystem with distributed locking and failover
> >> >
> >> > capability)
> >> >
> >> > > Regards,
> >> > > Luis
> >> > >
> >> > > */Sunil Mushran <Sunil.Mushran at oracle.com>/* wrote:
> >> > >
> >> > > Appears what you are looking for is a mix of ocfs2 and nfs.
> >> > > The storage servers mount the shared disks and the reexport
> >> > > them via nfs to the remaining servers.
> >> > >
> >> > > ubuntu 6.06 is too old. If you are stuck on Ubuntu LTS, the
> >> > > next version 7.10 should have all you want.
> >> > >
> >> > > Pavel Georgiev wrote:
> >> > > > Hi List,
> >> > > >
> >> > > > I`m trying to build a cluster storage with commodity
> >> >
> >> > hardware in
> >> >
> >> > > a way that
> >> > >
> >> > > > the all the data would be on > 1 server. It should
> >> >
> >> > have the meet
> >> >
> >> > > the
> >> > >
> >> > > > following requirements:
> >> > > > 1) If one of the servers goes down, the cluster
> >> >
> >> > should continue
> >> >
> >> > > to work with
> >> > >
> >> > > > rw access from all clients.
> >> > > > 2) Clients that mount the storage should not be part
> >> >
> >> > of cluster
> >> >
> >> > > (not export
> >> > >
> >> > > > any disk storage) - I have few servers with huge disks that I
> >> > >
> >> > > want to store
> >> > >
> >> > > > data on (currently 2 servers, maybe more in the future) and I
> >> > >
> >> > > want to
> >> > >
> >> > > > storethe data only on them, the rest of the server should just
> >> > >
> >> > > mount and use
> >> > >
> >> > > > that storage with the ability to continue operation if one of
> >> > >
> >> > > the two storage
> >> > >
> >> > > > servers goes down.
> >> > > > 3) More servers should be able to join the cluster
> >> >
> >> > and at given
> >> >
> >> > > point,
> >> > >
> >> > > > expanding the total size of the cluster, hopefully without
> >> > >
> >> > > rebuilding the
> >> > >
> >> > > > storage.
> >> > > > 4) Load balance is not a issue - all the load can go to one of
> >> > >
> >> > > the two storage
> >> > >
> >> > > > servers (although its better to be balanced), the main goal is
> >> > >
> >> > > to have
> >> > >
> >> > > > redundant storage
> >> > > >
> >> > > > Does ocfs2 meet these requirements? I read few howtos but none
> >> > >
> >> > > of them
> >> > >
> >> > > > mentioned my second requirement (only some of the servers to
> >> > >
> >> > > hold the data).
> >> > >
> >> > > > Are there any specific steps to do to accomplish (2) and (3)?
> >> > > >
> >> > > > I`m using Ubuntu 6.06 on x86.
> >> > > >
> >> > > > Thanks!
> >> > > >
> >> > > > _______________________________________________
> >> > > > Ocfs2-users mailing list
> >> > > > Ocfs2-users at oss.oracle.com
> >> > > > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >> > >
> >> > > _______________________________________________
> >> > > Ocfs2-users mailing list
> >> > > Ocfs2-users at oss.oracle.com
> >> > > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >> >
> >> > --------------------------------------------------------------
> >> > ----------
> >> >
> >> > > Take the Internet to Go: Yahoo!Go puts the Internet in your pocket:
> >> >
> >> > <http://us.rd.yahoo.com/evt=48253/*http://mobile.yahoo.com/go?
> >> > refer=1GNXIC>
> >> >
> >> > > mail, news, photos & more.
> >> >
> >> > --------------------------------------------------------------
> >> > ----------
> >> >
> >> > > _______________________________________________
> >> > > Ocfs2-users mailing list
> >> > > Ocfs2-users at oss.oracle.com
> >> > > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >> >
> >> > _______________________________________________
> >> > Ocfs2-users mailing list
> >> > Ocfs2-users at oss.oracle.com
> >> > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >>
> >> _______________________________________________
> >> Ocfs2-users mailing list
> >> Ocfs2-users at oss.oracle.com
> >> http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >>
> >>
> >> _______________________________________________
> >> Ocfs2-users mailing list
> >> Ocfs2-users at oss.oracle.com
> >> http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >
> > _______________________________________________
> > Ocfs2-users mailing list
> > Ocfs2-users at oss.oracle.com
> > http://oss.oracle.com/mailman/listinfo/ocfs2-users
More information about the Ocfs2-users
mailing list