diff --git a/doc/source/installation_guide/duplex.rst b/doc/source/installation_guide/duplex.rst index c9547a7ff..fc42756de 100644 --- a/doc/source/installation_guide/duplex.rst +++ b/doc/source/installation_guide/duplex.rst @@ -1200,6 +1200,293 @@ excessive data-sync time). Use 'fm alarm-list' to confirm status. | 2 | controller-1 | controller | unlocked | enabled | available | +----+--------------+-------------+----------------+-------------+--------------+ +----------------------------------- +Extending the Compute Node Capacity +----------------------------------- + +You can add up to four Compute Nodes to the All-in-One Duplex Deployment. + +************************** +Compute Hosts Installation +************************** + +After initializing and configuring the two controllers, you can add up +to four additional compute hosts. To add a host, do the following: + +^^^^^^^^^^^^^^^^^ +Initializing Host +^^^^^^^^^^^^^^^^^ + +Power on the host. The following appears in the host console: + +:: + + Waiting for this node to be configured. + + Please configure the personality for this node from the + controller node in order to proceed. + +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +Updating the Host Name and Personality +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + +On Controller-0, acquire Keystone administrative privileges: + +:: + + controller-0:~$ source /etc/nova/openrc + + +Wait for the Controller-0 to both discover the new host and to list that host +as UNKNOWN in the table: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-list + +----+--------------+-------------+----------------+-------------+--------------+ + | id | hostname | personality | administrative | operational | availability | + +----+--------------+-------------+----------------+-------------+--------------+ + | 1 | controller-0 | controller | unlocked | enabled | online | + | 2 | controller-1 | controller | unlocked | enabled | online | + | 3 | None | None | locked | disabled | offline | + +----+--------------+-------------+----------------+-------------+--------------+ + +Use the system host-update command to update the host personality attribute: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-update 3 personality=compute hostname=compute-0 + +See also: 'system help host-update' + +Unless it is known that the host's configuration can support the +installation of more than one node, it is recommended that the +installation and configuration of each node be serialized. For example, +if the virtual disks for the entire cluster are hosted on the host's +root disk and that disk happens to be a single rotational type hard disk, +then the host cannot reliably support parallel node installation. + + +^^^^^^^^^^^^^^^ +Monitoring Host +^^^^^^^^^^^^^^^ + +On Controller-0, you can monitor the installation progress by periodically +running the system host-show command for the host. Progress is +shown in the install_state field. + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-show | grep install + | install_output | text | + | install_state | booting | + | install_state_info | None | + + +Wait while the host is installed, configured, and rebooted. Depending on +hardware, it could take up to 20 minutes for this process to complete. +When the reboot is complete, the host is reported as Locked, Disabled, +and Online. + +^^^^^^^^^^^^^ +Listing Hosts +^^^^^^^^^^^^^ + +You can use the system host-list command to list the hosts once the node +has been installed, configured, and rebooted: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-list + +----+--------------+-------------+----------------+-------------+--------------+ + | id | hostname | personality | administrative | operational | availability | + +----+--------------+-------------+----------------+-------------+--------------+ + | 1 | controller-0 | controller | unlocked | enabled | online | + | 2 | controller-1 | controller | unlocked | enabled | online | + | 3 | compute-0 | compute | locked | disabled | online | + +----+--------------+-------------+----------------+-------------+--------------+ + + +********************** +Compute Host Provision +********************** + +You must configure the network interfaces and the storage disks on a +host before you can unlock it. For each Compute Host, do the following: + +On Controller-0, acquire Keystone administrative privileges: + +:: + + controller-0:~$ source /etc/nova/openrc + + +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +Provisioning Network Interfaces on a Compute Host +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + +In order to identify hardware port names, types, and discovered +pci-addresses on Controller-0, list the host ports: + +- **Only in Virtual Environment**: Ensure that the interface used is + one of those attached to the host bridge with model type "virtio" (i.e. + eth1000 and eth1001). The model type "e1000" emulated devices will + not work for provider networks: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-port-list compute-0 + +--------------------------------------+---------+----------+--------------+... + | uuid | name | type | pci address |... + +--------------------------------------+---------+----------+--------------+... + | de9ec830-cf33-4a06-8985-cd3262f6ecd3 | enp2s1 | ethernet | 0000:02:01.0 |... + | 9def88fb-d871-4542-83e2-d4b6e35169a1 | enp2s2 | ethernet | 0000:02:02.0 |... + | b2e38da9-840f-446c-b466-ceb6479266f6 | eth1000 | ethernet | 0000:02:03.0 |... + | c1694675-643d-4ba7-b821-cd147450112e | eth1001 | ethernet | 0000:02:04.0 |... + +--------------------------------------+---------+----------+--------------+... + + +Use the following command to provision the data interface for Compute: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-if-modify -p providernet-a -c data compute-0 eth1000 + + +^^^^^^^^^^^^^^^^^^^^^^^^^^^ +VSwitch Virtual Environment +^^^^^^^^^^^^^^^^^^^^^^^^^^^ + +**Only in Virtual Environment**. If the compute node has more than four CPUs, +the system auto-configures the vswitch to use two cores. However, some virtual +environments do not properly support multi-queue, which is required in a +multi-CPU environment. Therefore, run the following command to reduce the +vswitch cores to one: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-cpu-modify compute-0 -f vswitch -p0 1 + +--------------------------------------+-------+-----------+-------+--------+... + | uuid | log_c | processor | phy_c | thread |... + +--------------------------------------+-------+-----------+-------+--------+... + | 9d53c015-8dd5-4cd8-9abb-6f231fa773c2 | 0 | 0 | 0 | 0 |... + | 8620eaa7-19af-4ef7-a5aa-690ed57f01c7 | 1 | 0 | 1 | 0 |... + | f26598a5-8029-4d20-999e-2ec5d532d62e | 2 | 0 | 2 | 0 |... + | f7b3ce4a-10da-485d-873c-4e6988f7e9cf | 3 | 0 | 3 | 0 |... + +--------------------------------------+-------+-----------+-------+--------+... + + +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +Provisioning Storage on a Compute Host +^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + +Review the available disk space and capacity and then obtain the uuid(s) of +the physical disk(s) to be used for nova local: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-list compute-0 + +--------------------------------------+-----------+---------+---------+-------+------------+... + | uuid | device_no | device_ | device_ | size_ | available |... + | | de | num | type | gib | gib |... + +--------------------------------------+-----------+---------+---------+-------+------------+... + | 0ae45272-c9f4-4824-8405-f6c8946fda1e | /dev/sda | 2048 | HDD | 200.0 | 120.976 |... + | d751abfe-de57-4b23-b166-1d3d5b4d5ca6 | /dev/sdb | 2064 | HDD | 200.0 | 199.997 |... + +--------------------------------------+-----------+---------+---------+-------+------------+... + +Use the following command to create the 'nova-local' local volume group: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-lvg-add compute-0 nova-local + +-----------------------+-------------------------------------------------------------------+ + | Property | Value | + +-----------------------+-------------------------------------------------------------------+ + | lvm_vg_name | nova-local | + | vg_state | adding | + | uuid | 78e84a57-c40e-438c-97a4-49ebec6394d1 | + | ihost_uuid | 6f2de9b6-c55c-47b0-b40e-aff47f1e1bda | + | lvm_vg_access | None | + | lvm_max_lv | 0 | + | lvm_cur_lv | 0 | + | lvm_max_pv | 0 | + | lvm_cur_pv | 0 | + | lvm_vg_size_gib | 0.0 | + | lvm_vg_avail_size_gib | 0.0 | + | lvm_vg_total_pe | 0 | + | lvm_vg_free_pe | 0 | + | created_at | 2019-01-15T12:31:45.796538+00:00 | + | updated_at | None | + | parameters | {u'concurrent_disk_operations': 2, u'instance_backing': u'image'} | + +-----------------------+-------------------------------------------------------------------+ + + +Use the following command to create a disk partition to add to the volume +group based on the uuid of the physical disk: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-pv-add compute-0 nova-local d751abfe-de57-4b23-b166-1d3d5b4d5ca6 + +--------------------------+--------------------------------------------+ + | Property | Value | + +--------------------------+--------------------------------------------+ + | uuid | bba20f20-81fc-40a2-bf1e-a2fae849625b | + | pv_state | adding | + | pv_type | disk | + | disk_or_part_uuid | d751abfe-de57-4b23-b166-1d3d5b4d5ca6 | + | disk_or_part_device_node | /dev/sdb | + | disk_or_part_device_path | /dev/disk/by-path/pci-0000:00:1f.2-ata-2.0 | + | lvm_pv_name | /dev/sdb | + | lvm_vg_name | nova-local | + | lvm_pv_uuid | None | + | lvm_pv_size_gib | 0.0 | + | lvm_pe_total | 0 | + | lvm_pe_alloced | 0 | + | ihost_uuid | 6f2de9b6-c55c-47b0-b40e-aff47f1e1bda | + | created_at | 2019-01-15T12:32:40.535160+00:00 | + | updated_at | None | + +--------------------------+--------------------------------------------+ + + +^^^^^^^^^^^^^^^^^^^^^^^^ +Unlocking a Compute Host +^^^^^^^^^^^^^^^^^^^^^^^^ + +On Controller-0, use the system host-unlock command to unlock the +Compute node: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-unlock compute-0 + + +Wait while the Compute node is rebooted and re-configured. Depending on +hardware, it can take up to 10 minutes for the reboot to complete. Once +the reboot is complete, the nodes Availability State reports as "In-Test" +and is followed by unlocked/enabled. + +------------------- +System Health Check +------------------- + +*********************** +Listing StarlingX Nodes +*********************** + +On Controller-0, after a few minutes, all nodes are reported as +Unlocked, Enabled, and Available: + +:: + + [wrsroot@controller-0 ~(keystone_admin)]$ system host-list + +----+--------------+-------------+----------------+-------------+--------------+ + | id | hostname | personality | administrative | operational | availability | + +----+--------------+-------------+----------------+-------------+--------------+ + | 1 | controller-0 | controller | unlocked | enabled | available | + | 2 | controller-1 | controller | unlocked | enabled | available | + | 3 | compute-0 | compute | unlocked | enabled | available | + +----+--------------+-------------+----------------+-------------+--------------+ + ***************** System Alarm List