Hello, I'm Dean Pompilio and welcome to cyber ery virtual ization configuration installation and management course. Ron Module 11. Now Lesson number two. This lesson will be talking about configuring high availability. More specifically, how to configure high availability in a cluster.
So first of all, we think about what a cluster is.
This is just a collection of hosts and viens
that we can enable. High availability can also enable distributed resource scheduler.
We'll get into what distributed resource Scheduler is a little bit later in the course,
but basically DRS allows you to
automatically determined where the best places to put the EMS
based on the available resource is of your different hosts in the cluster.
It also allows for automatic or or some level of automated load balancing.
And for power management
you can even do. I distributed
resource scheduling for your power of your host themselves. It's kinda interesting concept,
So if we create a cluster, we get this interesting icon. It sort of looks a little bit like a BM icon, but they're taller, shapes their host shapes.
So this case you've got two hosts and to be EMS in this cluster.
talk a little bit about what some of the options are here.
if I want to turn on high availability, I right click by cluster in the inventory goto at its settings. There's a checkbox for enabling a change. There's also a checkbox for enabling. DRS.
Once you turn on H A,
then you have to think about some of the A J settings.
For instance, we have a checkbox for host monitoring.
This could be, ah, it's enabled by default, but it could be disabled if you're doing maintenance. For instance, if I was taking one of the hosts off line to do some work on it, maybe add more memory or something of that nature, I could disable monitoring once that monitoring is disabled. Now, when I take that host,
shut that hose down.
There will be no automatic movement of the EMS.
If you want to keep the PM's running, you would move them manually from one host to another two of the motion,
and that way you could take the second host down, and that wouldn't cause any problems. So this monitoring setting allows that
to either be on automatic behavior or something that you manually control.
Next, we have admission control.
So this, by default anyway, will disallow powering on of'em if it violates
the policies that you set up.
So if I have to host into a couple of the EMS,
I basically can decide what kind of resource is I need to have in reserve in order for the second VM, let's say to power up on the other host in the cluster.
That will make more sense when you see the lab and see how the sun is work.
But if emission controls enabled,
for instance, if hosted two goes away
and I want A and V M one was already running on host one. If VM two running out host one would violate the allowed amount of resource is, I'd like to keep in reserve
that when this is enabled, the VM won't be allowed to start.
If I disabled admission control now, it can let this be empower up on the other host, even if it does violate the policy I've chosen.
You might need to do that in some circumstances because that's only your only option to maintain up time.
But in general you might want to carefully consider leaving this option enabled so that you can properly control resource allocation between the hosts and your cluster.
As far as the admission control policy,
what we can do is set the number of host failures.
If I had a cluster of only two hosts and I could only tolerate a failure of one host
if I had a cluster with five hosts, I potentially could tolerate four host failures.
So that's kind of gives you an idea. And the default
setting would be one less than the number of hosts in your cluster.
But you could set that other ways to you could say, I've even if you had a five holes cluster you could only tolerate to host is failing,
so it's very configurable.
Then you have a percentage of resource is that you'd like to keep in reserve.
This is dealing with CPU and memory,
so going back to this example, if
if the VM one is running out, host one and it's using 60% of its Bram,
the M two wants to power up here and it wants to use 60% that would violate the policy. So going back to Mission Control. If the Michigan trolls enabled, the second BM would not be allowed to power on.
If I disabled mission control, then both PM's could try to use half of the memory on the host and because we can overcome it memory that might be possible
I can also specify fail over hosts specifically.
So let's say I've got ah, three host Cluster
and the 1st 2 hosts are the same. They're relatively low powered.
I might have 1/3 host,
which actually would appear here, but there's really no room to draw it. The third host might have a lot more memory and processing capacity than hosts one and two
S O. If there's a problem with Host one or two, I could specify always fail over a host three because it's got the most capacity.
So you've got a lot of really flexible options here for deciding how to deal with failures.
we have a little note here about slot size calculation. You'll see this in the lab as well.
look at a particular host and look at its resource allocation tab
or in other ways, you can see this, too. You can see how many slots
your host can support
or how many. How big the slots are,
and what this means is a given. VM
has a certain amount of CPU and memory overhead,
and ah, typical VM will use one slot or two slots or three slots, depending on what it's resource allocation requirements are.
So that's the way that the the resource is on your host are subdivided
so that the admission control and other H A algorithms can figure out where to move things around and what's possible. A ce faras, which hosts get which of'em is when there's a problem,
so we'll look at that in the lab. That'll make a little more sense when you see how that works and will change some settings and see how the slot size changes. Based on the resource allocation settings that we're modifying
now regarding the virtual machine configuration options,
this could be done at the cluster level or the VM level.
For instance, I can start with the restart priority of a virtual machine.
This could be disabled, which means that all the PM's will start at the same time
or I can pick low, medium or high priority on a pervy um basis,
and that's basically gives me a way to order the relative startup of the EMS. You might give high priority, for instance, to your most critical virtual machines, and then medium or low priority to those things that you don't need right away. But you still want to power up eventually.
If you get high priority to your critical, the EMS, then they get the most resource is and should boot the quickest.
So another way to adjust the prioritization of your boot time when a V M has to be restarted.
In the last lesson, we talked a little bit about the host isolation response.
If I've got a cluster built
and one of the hosts and the cluster loses its connection to the management network,
now we decide what has to happen here.
I could leave the host powered on.
If the mantra network is the only thing that's failed,
then leaving the host power down is a good idea because that means the PM's will continue to run
and we can maintain that status until the the management network gets restored. Maybe
cable got unplugged accidentally or network interface has
died on the switch, and maybe someone needs to move the two different switch port. These are things that might happen, right?
We could also power off the host if it gets isolated.
That's more of a drastic response, but it might make sense in certain circumstances to do that, or we can shut the host down.
It's more of a graceful shutdown,
and that might also make sense if the if the PM's don't get automatically moved somewhere, maybe what? We want to shut the host down and try to troubleshoot the problem. You bring it back up
so you get some flexibility there as well.
Then, within the cluster configuration screen, we also have another link for Vienna monitoring.
So we have some options. If VM were tools, if the heartbeat stops that Veum rituals provides,
then we can reset the BM.
Maybe maybe the more tools has crashed or stop for some other reason. Maybe someone accidentally shut it down
so we can restart the BM and that should bring up the tools when every boots
you can also adjust the monitoring sensitivity.
We're a little slider here from low to high
you might. You might want to adjust this and play around with different settings in your environment to find setting. That gives you the response you want without having the
The restarting of the EMS happened under circumstances that you don't wish it to happen.
Leaving in the middle might be a good place to begin before you start doing your testing.
Then within out of individual VM basis, I can adjust
how I want this BM to restart. I can use the cluster settings,
which means that I've got a global setting for all the PM's in that cluster.
Or you can go high, medium or low
for the priority for into individual VM or disable that.
So once you see the VM is running on your cluster in this configuration window would make sense is to right click
and select from the drop down menu. What your BM setting restart priority should be,
Then we have to think about some of the other the best practices here.
One thing to consider is having a redundant heartbeat network,
so the harpy gets sent between the master and slave hosts.
That way, they can each keep track of the other hosts on the cluster to know if that host is healthy and operational.
If ah host determines a determined to have a failure, this could be because no heartbeats are being received. The ping response is not working, and there's no heartbeat in the in the data store because you can your storage heartbeats as well.
In the lab, we use an NFS partition for that purpose.
So the heartbeat Network takes care of these heartbeats, signals and our packets network packets.
Oh, you do have to have a V M Colonel Port configured for their for management network. In order for this to work
and we'll see how that gets set up in the lab.
And if I can create a
a redundant harpy network now, I can have a even more reliable way to detect failures.
So a couple ways, you can do this. You can use Nick teeming
for the for the Heartbeat Network. Now I've got multiple physical interfaces in case I lose one of those. The other interfaces that remain can continue to send those heartbeats signals back and forth between all of the hosts,
or I can set up multiple Harper. The networks
and have a nick teeming and more redundant networks. I might have to nix for this harpy network one another to Nick's for Heartbeat Network. To Not I can tolerate three failures of those network cards and still maintain my heartbeat network.
All right, stay tuned for a part two of lesson to where we talk about configuring a J. Thank you.