Extending the CTERA Main Database or Catalog Node Storage Pool
Create a new storage block. This storage must be attached to the portal instance.
To extend the CTERA main database or catalog node storage pool:
1 For each CTERA Portal instance you have installed, stop CTERA services, by doing the following:
a Connect to the virtual server, over SSH, with the public IP, username and password.
b Run the following command to stop the CTERA Portal services:
ctera-portal-manage.sh stop
2 Extend the CTERA Storage Pool, by entering the following:
a Perform discovery on the storage.
# iscsiadm -m discovery -t sendtargets -p block_storage_ip
where block_storage_ip the target address IP for the new storage.
b Set the host to automatically log into the iSCSI array storage.
# iscsiadm -m node -L automatic
A successful message is returned.
c Verify that the host has logged in to the iSCSI array.
# iscsiadm -m session
# multipath -l
Note: With some Linux operating systems, the multipath -l command might return a warning that there is no host device. This warning can be ignored.
d Verify that the device is connected.
# fdisk -l | grep /dev/mapper
The disk is mounted and accessible to the host and a line similar to the following is returned:
Disk /dev/mapper//3600a09803830304e2f5d46447452774d: 43.0 GB, 42953867264 bytes, 83894272 sectors
Note: By default the device will attach to /dev/mapper/mpath_deviceID, where deviceID is the generated ID of the connected device.
e Create a partition on the disk: # fdisk /dev/mapper/disk_name
where disk_name is the value returned by the fdisk -l | grep /dev/mapper command, 3600a09803830304e2f5d46447452774d in the above example. Continue to create the partition, mostly using the defaults:
Command value
|
Meaning
|
n
|
Create a new partition.
|
p or Enter
|
Make the partition primary. p is the default.
|
1 or Enter
|
The partition number. 1 is the default.
|
Enter
|
The first sector or cylinder.
|
Enter
|
The last sector or cylinder.
|
t
|
Sets up the type of partition using the partition number specified.
|
8e
|
The Changes the type of partition from Linux to Linux LVM.
|
w
|
Writes the new partition information to the disk.
|
3 Reboot the virtual server for the disk changes to be recognized by the CTERA Portal.
4 List the storage to identify the name of the storage to create in the CTERA portal.
# ll /dev/mapper
A result similar to the following is returned:
lrwxrwxrwx 1 root root 7 Aug 23 08:42 36030304e2f5d464774a -> ../dm-0
lrwxrwxrwx 1 root root 7 Aug 23 08:42 36030304e2f5d464774a1 -> ../dm-2
lrwxrwxrwx 1 root root 7 Aug 23 08:42 36030304e2f5d464774d -> ../dm-1
lrwxrwxrwx 1 root root 7 Aug 23 08:42 36030304e2f5d464774d1 -> ../dm-4
crw------- 1 root root 10, 236 Aug 23 08:41 control
lrwxrwxrwx 1 root root 7 Aug 23 08:42 DataPool-DataPool -> ../dm-3
where dm-4 is the added storage.
5 Stop the portal server: # ctera-portal-manage.sh stop
6 Expose the added disk to the file system:
# ctera-storage-util.sh attach_endurance_disk device_ip [user pass iqn]
Where the following are taken from the Block Storage Details screen:
device_ip is the Target Address IP.
user is the Username.
pass is the Password. The password is not encrypted.
iqn is the Host IQN.
The name and address of the attached storage device is output with output similar to the following: /dev/mapper/3600a098038303051793f464462534b56.
7 Extend the storage in the portal: # ctera-storage-util.sh extend_storage /dev/dm-n
where dm-n is the name of the storage.
The main database or catalog node storage pool is extended.
8 start the portal server: # ctera-portal-manage.sh start
The portal starts.