[ClusterLabs] Antw: [EXT] Re: Why my node1 couldn't back to the clustering chain?
Ulrich Windl
Ulrich.Windl at rz.uni-regensburg.de
Fri Apr 9 05:06:14 EDT 2021
>>> Jason Long <hack3rcon at yahoo.com> schrieb am 09.04.2021 um 08:58 in Nachricht
<2055279672.56029.1617951519012 at mail.yahoo.com>:
> Thank you so much for your great answers.
> As the final questions:
> 1- Which commands are useful to monitoring and managing my pacemaker
> cluster?
My favorite is "crm_mon -1Arfj".
>
> 2- I don't know if this is a right question or not. Consider 100 PCs that
> each of them have an Intel Core 2 Duo Processor (2 cores) with 4GB of RAM.
> How can I merge these PCs together so that I have a system with 200 CPUs and
> 400GB of RAM?
If you don't just want to recycle old hardware, you could consider buying _one_ recent machine that has almost all that cores and RAM in one machine, probably saving a lot of power and space, too.
Like here:
# grep MHz /proc/cpuinfo | wc -l
144
# lscpu
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
Address sizes: 46 bits physical, 48 bits virtual
CPU(s): 144
On-line CPU(s) list: 0-143
Thread(s) per core: 2
Core(s) per socket: 18
Socket(s): 4
NUMA node(s): 4
Vendor ID: GenuineIntel
CPU family: 6
Model: 85
Model name: Intel(R) Xeon(R) Gold 6240 CPU @ 2.60GHz
Stepping: 7
CPU MHz: 1001.007
...
# free -h
total used free shared buff/cache available
Mem: 754Gi 1.7Gi 744Gi 75Mi 8.1Gi 748Gi
Regards,
Ulrich
>
>
>
>
>
>
> On Friday, April 9, 2021, 12:13:45 AM GMT+4:30, Antony Stone
> <antony.stone at ha.open.source.it> wrote:
>
>
>
>
>
> On Thursday 08 April 2021 at 21:33:48, Jason Long wrote:
>
>> Yes, I just wanted to know. In clustering, when a node is down and
>> go online again, then the cluster will not use it again until another node
>> fails. Am I right?
>
> Think of it like this:
>
> You can have as many nodes in your cluster as you think you need, and I'm
> going to assume that you only need the resources running on one node at any
> given time.
>
> Cluster management (eg: corosync / pacemaker) will ensure that the resources
>
> are running on *a* node.
>
> The resources will be moved *away* from that node if they can't run there
> any
> more, for some reason (the node going down is a good reason).
>
> However, there is almost never any concept of the resources being moved *to*
> a
> (specific) node. If they get moved away from one node, then obviously they
> need to be moved to another one, but the move happens because the resources
> have to be moved *away* from the first node, not because the cluster thinks
> they need to be moved *to* the second node.
>
> So, if a node is running its resources quite happily, it doesn't matter what
>
> happens to all the other nodes (provided quorum remains); the resources will
>
> stay running on that same node all the time.
>
>
> Antony.
>
> --
> Was ist braun, liegt ins Gras, und raucht?
> Ein Kaminchen...
>
>
> Please reply to the list;
> please *don't* CC
> me.
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
More information about the Users
mailing list