Building A Two Node SLES11 SP2 Linux Cluster On VMware

So… interested in building a linux cluster huh? If so you definitely have a few choices in OS. The big names (and what I would call the most reliable names) are Redhat Linux (a.k.a. CentOS a.k.a. Oracle) and Suse Linux Enterprise Server (a.k.a. SLES).

I’ve just completed my ebook giving you detailed step-by-step instructions on how to build your cluster. In this post I’ll give you the general outline as well as instructions how to download my ebook. Read on!

building-a-two-node-sles-cluster-fullsmallcover

“Building A Two Node SLES11 SP2 Linux Cluster On VMware”

I have a very super detailed step-by-step ebook on this very subject which I’m selling for ONLY $12.99! Interested? Get the Ebooks Now

Here is an outline of the steps on how to build your cluster.

This is a complete list of steps on how to build your cluster. If you need more detail on each step, I encourage you to check out my book.

  1. Install your two SLES 11 SP2 linux servers. Straightforward for the most part. I usually add the development package and take out firewalls (Novell AppArmor)
  2. Copy the .iso to local drive and add to software repository
    Especially if you do installs remotely, much nicer to use the .iso than to have people insert and eject discs. You can add them in Yast -> Software Management -> Software Repositories.

  3. Add two shared drives, one for quorum, one for data
    For physical servers, this usually means fiber storage through a switch or direct connection. For VMware, there are a few things you’ll need to do to add disks to VMs and have them be shared

  4. Add multipath support to your nodes
    Not required, but very handy for cluster builds

    (The above steps are covered in detail in my free excerpt of my ebook. Skip Down To My Ebooks Now)

  5. Install your cluster software and scripts
    This will be on the SLES HA media (not the base OS install media). Once you have added the disc or .iso as a repository you will see the “High Availability” option in yast. On the left side choose “High Availability” in the “Groups” search and install everything.

  6. Add your multicast cluster config info
    You will see a sample in the corosync.conf.example which is installed with the cluster software

  7. Start and verify your cluster
    With both nodes having the same multicast heartbeat, when you start the cluster you should be able to see both nodes. The cluster is started/stopped with the “openais” service. You can verify by looking in /var/log/messages for the string “TOTEM”. Also with the “ps” command you should now see the corosync process and the children heartbeat processes. You can also start using the crm commands now to take a look at your cluster:

    • crm configure show
    • crm_mon
    • crm_verify -L
    • Disable stonith and quorum
      Since you have not set up a quorum disk yet you want to tell your cluster to disable these for now

      • crm configure property stonith-enabled=false
      • crm configure property no-quorum-policy=ignore
    • In yast, configure and enable csync2
      You will want to make sure you do these things:

      • Auto Generate Node ID
      • Enable Security Auth
      • Generate Auth keyfile
      • Enable mgmtd
      • Add hosts and files to Csync2
      • Generate pre-shared keys
      • Turn Csync2 on
    • Copy over your Csync2 files to second node
      Make sure your initial config files are the same on both nodes. These files are the ones in the “Add Suggested Files” portion of your Csync2 config in yast.

    • Turn on the Csync2 service (part of inetd)
      You will need to enable Csycn2 and xinetd using chkconfig

    • Run your sync using “csync2 -xv”
      If it finishes with errors, check the files marked dirty, copy them manually, then run again

    • Start crm_gui (via vncserver session)
      It is much easier to configure and manage the cluster from here on out with the GUI. If you are remote and have a windows as your personal computer (like me), you will need to use VNC or else cygwin-x (I have a cygwin-x tutorial here)

        You’re ALMOST THERE!!

        • Set a password for “hacluster” and start the crm_gui and log in with the hacluster user/pass
        • Now you can start adding resources. First add a resource group
          A resource group is what is failed over from node to node. Then you will put your resources in that group, i.e. all the things you want to failover together

        • Add a IP address primitive resource to the resource group
        • At this point you can start your resource group and you will see the cluster IP on the active node
          Now you can failover the resource from node to node using reboots or anything else to take a node out of the cluster. I always do this as my initial test because an IP address does not depending on anything else being correct. In other words, with shared storage, there are a lot of variables. So to make sure my cluster is working, I use just the IP address

        • Add your shared data disk now as a resource
          The same way you added a primitive resource IP to your resource group, you can now add a filesystem/device

        • Add your quorum disk
          First you use the command “sbd -d /dev/[path to quorum disk] create” to make your device usable as a quorum disk. Next you create the file “/etc/sysconfig/sbd” with these contents

          SBD_DEVICE="/dev/mapper/sbddisk"
          # The next line enables the watchdog support:
          SBD_OPTS="-W"

          Now you restart your openais service (there will be an error upon stopping it, but it should start fine

        • Add your stonith disk to your cluster in the crm_gui
          This time you add a primitve resource of type “Stonith” and add your stonith a.k.a. quorum disk. Do not add it to your resource group, it is a resource on its own separate from the ip/shared data resource group.

        • Use command line to add “stonith-enabled” back into your cluster config
          This one I haven’t found the GUI version so we do command line again (remember we took it out earlier?). Use the command “crm configure property stonith-enabled=true”

        • Back in crm_gui, you can now start your stonith resource

            SERIOUSLY… ALMOST DONE!

            • Create an application start script
              It needs to be LSB compliant and answer to

              • start
              • stop
              • restart
              • reload
              • force-reload
              • status

              The easiest way is to take an existing script in /etc/init.d and modify it for your own liking. Put this script in the directory /etc/ha.d/resource.d

            • In crm_gui, add the application to your resource group
              Now you can add an application to your resource group. It’s a primitve type resource again.

            DONE!!

            Your cluster should now have a primitive resource quorum disk and a resource group containing an ip address, shared data disk, and application. This is a very basic two node cluster but models well what you will do in real life. For example building Apache Linux clusters or Oracle Linux clusters. Those are at the very basic applications with shared disk you want to failover from one node to the other.

            Need more help?

            Want to look over my shoulder as I build my cluster?

            Well, you have two choices available to you:

            I have written a 130 page ebook chock full of step-by-step instructions with screenshots and sample screen output. If you want me to hold your hand while you build your cluster, this ebook will do just that with detailed step-by-step-by-step instructions. And right after you pay (via paypal) you will get immediate download access

            LOVE IT! For only $12.99, I’LL BUY IT NOW!

              There’s even a free preview. I have the first 40 pages available for download for free so you can check out the detail I go into. Most of the first 40 pages are strictly setting up your VMware nodes for clustering and then setting up multipath. Should give you a good idea of the detail I’ll go into.


              Want to purchase or download the excerpt? Details below!

                Building A Two Node SLES11 SP2 Linux Cluster On VMware (FULL VERSION) $12.99
                building-a-two-node-sles-cluster-fullsmallcover

                130 pages of detailed step-by-step instruction to build a two node SUSE Linux Enterprise Server Linux Cluster on VMware environment. Chock full of screenshots taking you from the beginning to the setting up a two node cluster with failover IP address, shared disk, and application. To view a sample, you can read the first 40 pages for free.

                Building A Two Node SLES11 SP2 Linux Cluster On VMware (EXCERPT) FREE
                building-a-two-node-sles-cluster-excerpt-smallcover
                Download Now

                40 page excerpt of my 130 page book on how to build a Suse Linux Enterprise Linux Cluster on ESXi VMWare environment. Chock full of screenshots and step-by-step instructions. The full version is 130 pages and for sale for $12.99 in the “Paid Ebooks” section below.

                CHEERS! Happy Cluster Building!

                  Geek, Linux, VMware

                  5 thoughts on “Building A Two Node SLES11 SP2 Linux Cluster On VMware

                      • Hi all, thanks for your comments.

                        Savio: download the free sample doc. It has the table of contents and also the detail I go into. A few notes: I am using the SLES bundled HA media (which is pacemaker/openais). You can see by the table of contents I will have built a two node cluster with shared IP, shared filesystem, and shared application.

                  Leave a Comment

                  Your email address will not be published. Required fields are marked *

                  Spam protection by WP Captcha-Free