2
0
mirror of https://github.com/xcat2/xcat-core.git synced 2025-05-21 11:12:04 +00:00

Refactor the entire "Advanced" section of the documentation

- categorize into topics without "xCAT" in the heading
  since the topics are all related to xCAT
- Alphabetize the topics to make it easier to find
- create sub directories for the topics to contain
  the documentation
- broke down some of the docs to multiple files to help
  ease of readability
This commit is contained in:
Victor Hu 2015-10-07 09:55:33 -04:00
parent 12127eb09b
commit bb279bd76f
50 changed files with 193 additions and 50 deletions

View File

@ -1,2 +0,0 @@
Changing the Hostname or IP for Management Node
===============================================

View File

@ -1,2 +0,0 @@
Chaning the Hostname or IP for Compute Node
===========================================

View File

@ -1,2 +1,2 @@
Configuring IPv6 in xCAT Cluster
Changing the hostname/IP address
================================

View File

@ -0,0 +1,7 @@
Compute Node
============
.. toctree::
:maxdepth: 2
changing_hostname_ip.rst

View File

@ -0,0 +1,9 @@
Cluster Maintenance
===================
.. toctree::
:maxdepth: 2
compute_node/index.rst
mgmt_node/index.rst
sw_fw_inventory.rst

View File

@ -0,0 +1,4 @@
Changing the hostname/IP address
================================
There may be times when it is required to change the hostname or IP address of the xCAT management nodes. This document helps to outline the steps required to ensure successfully changing this attribute on the management node.

View File

@ -0,0 +1,7 @@
Management Node
===============
.. toctree::
:maxdepth: 2
changing_hostname_ip.rst

View File

@ -0,0 +1,7 @@
GPUs
====
.. toctree::
:maxdepth: 2
cuda.rst

View File

@ -1,22 +1,17 @@
========
Overview
========
The xCAT management node plays an important role in the cluster, if the management node is down for whatever reason, the administrators will lose the management capability for the whole cluster, until the management node is back up and running. In some configuration, like the Linux nfs-based statelite in a non-hierarchy cluster, the compute nodes may not be able to run at all without the management node. So, it is important to consider the high availability for the management node.
The goal of the HAMN(High Availability Management Node) configuration is, when the primary xCAT management node fails, the standby management node can take over the role of the management node, either through automatic failover or through manual procedure performed by the administrator, and thus avoid long periods of time during which your cluster does not have active cluster management function available.
============================
Configuration considerations
============================
xCAT provides several configuration options for the HAMN, you can select one of the option based on your failover requirements and hardware configuration, the following configuration considerations should be able to help you to make the decision.
******************************
Data synchronization mechanism
******************************
------------------------------
The data synchronization is important for any high availability configuration. When the xCAT management node failover occurs, the xCAT data needs to be exactly the same before failover, and some of the operating system configuration should also be synchronized between the two management nodes. To be specific, the following data should be synchronized between the two management nodes to make the xCAT HAMN work:
@ -35,9 +30,8 @@ Warning: Running database through network file system has a lot of potential pro
**3\. Mirroring**: each of the management node has its own copy of the xCAT data, and the two copies of data are syncronized through mirroring mechanism. DRBD is used widely in the high availability configuration scenarios, to provide data replication by mirroring a whole block device via network. If we put all the important data for xCAT onto the DRBD devices, then it could assure the data is synchronized between the two management nodes. Some parallel file system also provides capability to mirror data through network.
*************************************
Manual failover VS automatic failover
*************************************
Manual vs. Automatic Failover
-----------------------------
When the primary management node fails, the backup management node could automatically take over, or the administrator has to perform some manual procedure to finish the failover. In general, the automatic failover takes less time to detect the failure and perform and failover, comparing with the manual failover, but the automatic failover requires more complex configuration. We could not say the automatic failover is better than the manual failover in all cases, the following factors should be considered when deciding the manual failover or automatic failover:
@ -55,7 +49,6 @@ The configuration for the high availability applications is usually complex, it
The automatic failover brings in several high availability applications, after the initial configuration is done, additional maintenance effort will be needed. For example, taking care of the high availability applications during cluster update, the updates for the high availability applications themselves, trouble shooting any problems with the high availability applications. A simple question may be able to help you to decide: could you get technical support if some of the high availability applications run into problems? All software has bugs ...
=====================
Configuration Options
=====================

View File

@ -1,5 +1,7 @@
Highly Available Management Node
================================
High Avaiability
================
The following pages describes ways to configure the xCAT Management Node for High Availbility.
.. toctree::
:maxdepth: 2

View File

@ -0,0 +1,7 @@
Hierarchical Clusters
=====================
.. toctree::
:maxdepth: 2
introduction.rst

View File

@ -0,0 +1,10 @@
Introduction
============
In supporting large clusters, it is desirable to have more than a single management node handling the installation and management of compute nodes.
In xCAT, these additional nodes are referred to as *Service Nodes (SN)*. The management node can delegate all management operations for a compute node to the Service node that is managing them. You can have one of more service nodes configured to install and manage a group of compute nodes.
Service Nodes
-------------

View File

@ -1,2 +0,0 @@
Managing the Mellanox Infiniband Network
========================================

View File

@ -4,24 +4,22 @@ Advanced Topics
.. toctree::
:maxdepth: 2
restapi/index.rst
mixed_cluster/index.rst
hierarchy_cluster.rst
confluent/index.rst
cluster_maintenance/index.rst
docker/index.rst
gpu/index.rst
hamn/index.rst
hierarchy/index.rst
mixed_cluster/index.rst
networks/index.rst
change_mn_ip.rst
change_node_ip.rst
xcat_security.rst
ports/xcat_ports.rst
raid/index.rst
security/index.rst
softlayer/index.rst
kit/index.rst
ib.rst
confluent/index.rst
docker/index.rst
ipv6_support.rst
xcat_port.rst
sysclone.rst
switch_mgmt.rst
vlan.rst
zone.rst
softlayer.rst
firmware_inventory.rst
configraid.rst
cuda.rst
switches/index.rst
sysclone/index.rst
webservices/index.rst
zones/index.rst

View File

@ -0,0 +1,9 @@
Networks
========
.. toctree::
:maxdepth: 2
infiniband/index.rst
ipv6/index.rst
vlan/index.rst

View File

@ -0,0 +1,2 @@
IB Driver Preparation and Installation
======================================

View File

@ -0,0 +1,2 @@
Firmware Updates
================

View File

@ -0,0 +1,11 @@
Infiniband (Mellanox)
=====================
.. toctree::
:maxdepth: 2
driver_and_installation.rst
network_configuration.rst
switch_configuration.rst
ufm_configuration.rst
firmware_updates.rst

View File

@ -0,0 +1,2 @@
IB Network Configuration
========================

View File

@ -0,0 +1,2 @@
IB Switch Configuration
=======================

View File

@ -0,0 +1,2 @@
UFM Configuration
=================

View File

@ -0,0 +1,7 @@
IPv6
====
.. toctree::
:maxdepth: 2
ipv6_support.rst

View File

@ -0,0 +1,2 @@
Configuring IPv6
================

View File

@ -0,0 +1,7 @@
VLANs
=====
.. toctree::
:maxdepth: 2
vlan.rst

View File

@ -0,0 +1,4 @@
Port Usage
==========
TODO: http://sourceforge.net/p/xcat/wiki/XCAT_Port_Usage/?version=4

View File

@ -1,5 +1,5 @@
Configuring hardware RAID in xCAT cluster
=========================================
Hardware RAID
=============
Overview
--------

View File

@ -0,0 +1,7 @@
RAID
====
.. toctree::
:maxdepth: 2
hardware_raid.rst

View File

@ -1,2 +0,0 @@
REST API Reference
==================

View File

@ -0,0 +1,8 @@
Security
========
TODO, organize/rewrite from http://sourceforge.net/p/xcat/wiki/XCAT_2_Security/
.. toctree::
:maxdepth: 2

View File

@ -0,0 +1,7 @@
SoftLayer
=========
.. toctree::
:maxdepth: 2
softlayer.rst

View File

@ -1,2 +0,0 @@
Managing Ethernet Switches
==========================

View File

@ -0,0 +1,2 @@
Ethernet Switches
=================

View File

@ -0,0 +1,7 @@
Switch Management
=================
.. toctree::
:maxdepth: 2
ethernet_switches.rst

View File

@ -0,0 +1,7 @@
System Clone (sysclone)
=======================
.. toctree::
:maxdepth: 2
sysclone.rst

View File

@ -1,5 +1,5 @@
xCAT Web Services REST API
==========================
Web Services
============
.. toctree::
:maxdepth: 2

View File

@ -0,0 +1,2 @@
RESTful APIs
============

View File

@ -1,7 +1,7 @@
Set Up Web Service for REST API
===============================
Configure Web Services
======================
The following steps describe how to setup the WEB Service to use the REST API
The following steps describe how to setup the Web Service to use the REST API
Enable the HTTPS service for REST API
-------------------------------------

View File

@ -1,2 +0,0 @@
xCAT 2 Security
===============

View File

@ -0,0 +1,2 @@
Change Zones
============

View File

@ -1,2 +1,2 @@
xCAT Port Usage
Configure Zones
===============

View File

@ -0,0 +1,2 @@
Create Zones
============

View File

@ -0,0 +1,11 @@
Zones
=====
.. toctree::
:maxdepth: 2
overview.rst
configure_zones.rst
create_zones.rst
change_zones.rst
remove_zones.rst

View File

@ -0,0 +1,4 @@
Overview
========
xCAT supports the concept of zones within a single xCAT cluster managed by one (1) Management Node. The nodes in the cluster can be divided up into multiple zones that have different ssh keys managed separately.

View File

@ -0,0 +1,2 @@
Remove Zones
============