Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Prepare Cell Member for Federation

Prepare mount points

The cell member is usually "bare metal" server with NVMe disks for maximizing server throughput.

...

Store Nodemount point pathDisk BayDisk SerialDisk PathCPU No
2/data2



2/data22



3/data3



3/data33



4/data4



4/data44




How to identify Disk OS path and Disk serial
  1. To identify which of the server's NVMe disk bay is bound to which of the CPU use the hardware manufacture documentation.
    Also, write down the disk's serial number by visually observing the disk.
  2. In order to identify the disk os path (example : /dev/nvme01n) and the disk serial the user should install the NVMe disk utility software provided by the hardware supplier. Example : for Intel based NVMe SSD disks install the "Intel® SSD Data Center Tool"  (isdct).
    Example output of the Intel SSD DC tool :

    Code Block
    themeRDark
    isdct  show -intelssd
    
    - Intel SSD DC P4500 Series PHLE822101AN3PXXXX -
    
    Bootloader : 0133
    DevicePath : /dev/nvme0n1
    DeviceStatus : Healthy
    Firmware : QDV1LV45
    FirmwareUpdateAvailable : Please contact your Intel representative about firmware update for this drive.
    Index : 0
    ModelNumber : SSDPE2KE032T7L
    ProductFamily : Intel SSD DC P4500 Series
    SerialNumber : PHLE822101AN3PXXXX
    
    


  3. Use the disks bay number and the disk serial number (visually identified) and correlate with the output of the disk tool to identify the disk os path.


Examples for Mount Points and Disk Configurations


Store Nodemount point pathDisk BayDisk SerialDisk PathCPU No
2/data21PHLE822101AN3PXXXX/dev/nvme0n11
2/data222
/dev/nvme1n11
3/data34
/dev/nvme2n12
3/data335
/dev/nvme3n12
4/data412
/dev/nvme4n13
4/data4413
/dev/nvme5n13


Example for LVM Configuration
Code Block
themeRDark
pvcreate -ff /dev/nvme0n1
vgcreate vg_data2 /dev/nvme0n1
lvcreate -l 100%FREE -n lv_data vg_data2
mkfs.xfs -f /dev/vg_data2/lv_data

pvcreate -ff /dev/nvme1n1
vgcreate vg_data22 /dev/nvme1n1
lvcreate -l 100%FREE -n lv_data vg_data22
mkfs.xfs /dev/vg_data22/lv_data

...

Code Block
themeRDark
/dev/vg_data2/lv_data    /data2                   xfs     defaults        0 0
/dev/vg_data22/lv_data   /data22                   xfs     defaults        0 0
/dev/vg_data3/lv_data    /data3                   xfs     defaults        0 0
/dev/vg_data33/lv_data   /data33                   xfs     defaults        0 0
/dev/vg_data4/lv_data    /data4                   xfs     defaults        0 0
/dev/vg_data44/lv_data   /data44                   xfs     defaults        0 0
Example for the final configuration for 3 Store's nodes
Note

Not including other mount points needed as describe on DPOD Hardware and Software Requirements

...

Code Block
themeRDark
# lsblk

NAME                MAJ:MIN RM   SIZE RO TYPE MOUNTPOINT
nvme0n1             259:0    0   2.9T  0 disk
└─vg_data2-lv_data  253:6    0   2.9T  0 lvm  /data2
nvme1n1             259:5    0   2.9T  0 disk
└─vg_data22-lv_data 253:3    0   2.9T  0 lvm  /data22
nvme2n1             259:1    0   2.9T  0 disk
└─vg_data3-lv_data  253:2    0   2.9T  0 lvm  /data3
nvme3n1             259:2    0   2.9T  0 disk
└─vg_data33-lv_data 253:5    0   2.9T  0 lvm  /data33
nvme4n1             259:4    0   2.9T  0 disk
└─vg_data44-lv_data 253:7    0   2.9T  0 lvm  /data44
nvme5n1             259:3    0   2.9T  0 disk
└─vg_data4-lv_data  253:8    0   2.9T  0 lvm  /data4

Cell Member Federation

...

Prepare local OS based firewall

Most Linux based OS uses local firewall service (iptables / firewalld).

The OS for "non Appliance Mode" DPOD installation is provided by the user and its the user's responsibility to allow needed connectivity to and from the server.

User should make sure needed connectivity detailed on table 1 is allowed on the OS local firewall service.

Note

When using DPOD "Appliance" mode installation for the cell manager, local OS based firewall service is handled by the cell member federation script.


Cell Member Federation

In order to federate and configure the cell member run the following script on in the cell manager once per cell member - e.g. if you want to add twocell members, run the script twice (in the cell manager), first time with the IP address of the first cell member, and second time with the IP address of the second cell manager.

...

  1. The following communication and ports are used in a remote collector deployment scenario (table 1). Perform the following commands to accomplish this task on each DPOD local firewall:

    Run in Local Node -
    Change the XXXX to the IP of the Remote Collector

    Code Block
    iptables -I INPUT -p tcp -s XXXX/24 --dport 9300:9309 -j ACCEPT
    service iptables save
    service iptables restart
    


    After running the commands, run the following command and search the output for two entries showing port 9300 (shown in red in the below screenshot)

    Code Block
    iptables -L -n


    Anchortable1table1table 1

    From

    To

    Ports (Defaults)

    Protocol

    Usage

    Local Node DPOD Appliance  

    Each Monitored Device

    5550 (TCP)

    HTTP/S

    Monitored Device administration management interface

    Local Node DPOD Appliance 

    DNS Server

    TCP and UDP 53

    DNS

    DNS services

    Local Node DPOD Appliance  

    NTP Server

    123 (UDP)

    NTP

    Time synchronization

    Local Node DPOD Appliance  

    Organizational mail server

    25 (TCP)

    SMTP

    Send reports by email

    Local Node DPOD Appliance  

    LDAP

    TCP 389 / 636 (SSL).

    TCP 3268 / 3269 (SSL)

    LDAP

    Authentication & authorization. Can be over SSL

    NTP Server

    Local Node DPOD Appliance  

    123 (UDP)

    NTP

    Time synchronization

    Each Monitored Device

    Local Node DPOD Appliance  

    60000-60009 (TCP)

    TCP

    SYSLOG Data

    Each Monitored Device

    Local Node DPOD Appliance  

    60020-60029 (TCP)

    HTTP/S

    WS-M Payloads

    FROM Users IPs         

    Local Node DPOD Appliance  

    443 (TCP)

    HTTP/S

    Access to with IBM DataPower Operations Dashboard Console

    FROM Admins IPs         

    Local Node DPOD Appliance  

    22 (TCP)

    TCP

    SSH

    Remote Collector DPOD Appliance  

    Each Monitored Device

    5550 (TCP)

    HTTP/S

    Monitored Device administration management interface

    Remote Collector DPOD Appliance 

    DNS Server

    TCP and UDP 53

    DNS

    DNS services

    Remote Collector DPOD Appliance  

    NTP Server

    123 (UDP)

    NTP

    Time synchronization

    Remote Collector DPOD Appliance  

    Organizational mail server

    25 (TCP)

    SMTP

    Send reports by email

    Remote Collector DPOD Appliance  

    LDAP

    TCP 389 / 636 (SSL).

    TCP 3268 / 3269 (SSL)

    LDAP

    Authentication & authorization. Can be over SSL

    NTP Server

    Remote Collector DPOD Appliance  

    123 (UDP)

    NTP

    Time synchronization

    Each Monitored Device

    Remote Collector DPOD Appliance  

    60000-60009 (TCP)

    TCP

    SYSLOG Data

    Each Monitored Device

    Remote Collector DPOD Appliance  

    60020-60029 (TCP)

    HTTP/S

    WS-M Payloads

    FROM Users IPs         

    Remote Collector DPOD Appliance  

    443 (TCP)

    HTTP/S

    Access to with IBM DataPower Operations Dashboard Console

    FROM Admins IPs         

    Remote Collector DPOD Appliance  

    22 (TCP)

    TCP

    SSH

    From the Local Node's UI, go to the Manage menu, select "Nodes" under "System" and click "Edit"
    Image Removed
    Enter the IP address of the Remote Collector deviceand click "Update", you can leave the "Agents DNS Address" empty 
    Image Removed
     In the Local Node
    Connect to the Local Node DPOD via ssh as root user (using putty or any other ssh client)
    Using the Command Line Interface choose option 2 - "Stop All", and wait until all the services are stopped, this may take a few minutes to complete.
    Image Removed

    In the Local Node
    Using putty or any other ssh client, issue the following command: 

    Code Block
    sed -i -e "s/^SERVICES_SIXTH_GROUP=\".*MonTier-SyslogAgent-1 MonTier-HK-WdpServiceResources MonTier-HK-WdpDeviceResources/SERVICES_SIXTH_GROUP=\"MonTier-HK-WdpServiceResources MonTier-HK-WdpDeviceResources/g" /etc/sysconfig/MonTier

    In the Local Node
    Using putty or any other ssh client, issue the following command: 

    Code Block
    mv /etc/init.d/MonTier-SyslogAgent-1 /etc/init.d/Disabled-MonTier-SyslogAgent-1
    mv /etc/init.d/MonTier-SyslogAgent-2 /etc/init.d/Disabled-MonTier-SyslogAgent-2
    mv /etc/init.d/MonTier-SyslogAgent-3 /etc/init.d/Disabled-MonTier-SyslogAgent-3
    mv /etc/init.d/MonTier-SyslogAgent-4 /etc/init.d/Disabled-MonTier-SyslogAgent-4
    mv /etc/init.d/MonTier-SyslogAgent-5 /etc/init.d/Disabled-MonTier-SyslogAgent-5
    mv /etc/init.d/MonTier-SyslogAgent-6 /etc/init.d/Disabled-MonTier-SyslogAgent-6
    mv /etc/init.d/MonTier-SyslogAgent-7 /etc/init.d/Disabled-MonTier-SyslogAgent-7
    mv /etc/init.d/MonTier-SyslogAgent-8 /etc/init.d/Disabled-MonTier-SyslogAgent-8
    mv /etc/init.d/MonTier-SyslogAgent-9 /etc/init.d/Disabled-MonTier-SyslogAgent-9
    mv /etc/init.d/MonTier-SyslogAgent-10 /etc/init.d/Disabled-MonTier-SyslogAgent-10
    
    
    mv /etc/init.d/MonTier-WsmAgent-1 /etc/init.d/Disabled-MonTier-WsmAgent-1
    mv /etc/init.d/MonTier-WsmAgent-2 /etc/init.d/Disabled-MonTier-WsmAgent-2
    mv /etc/init.d/MonTier-WsmAgent-3 /etc/init.d/Disabled-MonTier-WsmAgent-3
    mv /etc/init.d/MonTier-WsmAgent-4 /etc/init.d/Disabled-MonTier-WsmAgent-4
    mv /etc/init.d/MonTier-WsmAgent-5 /etc/init.d/Disabled-MonTier-WsmAgent-5
    
    Note: some errors might appear for services that are not exists in your specific deployment architecture type - for example "mv: cannot stat ‘/etc/init.d/Disabled-MonTier-SyslogAgent-10’: No such file or directory"
     In the Local Node
    Using any text editor (like vi), edit /etc/hosts files (e.g. vi /etc/hosts)
    Change the following entries:
    montier-es from 127.0.0.1 to the IP of the Local node device
    montier-syslog and montier-wsm to the IP of the remote collector device
    Image Removed
    you should save the changes when exit (e.g wq)
    In the Local Node
    Using the Command Line Interface -  Select option 1 "Start All", this may take a few minutes to complete
    Image Removed
    Connect to the Remote Collector DPOD via ssh as root user (using putty or any other ssh client)
    Using the Command Line Interface choose option 2 - "Stop All", and wait until all the services are stopped, this may take a few minutes to complete.
    Image Removed

    In the Remote Collector
    Using putty or any other ssh client, issue the following commands:

    Code Block
    mv /etc/init.d/MonTier-es-raw-trans-Node-1 /etc/init.d/Disabled-MonTier-es-raw-trans-Node-1
    mv /etc/init.d/MonTier-es-raw-trans-Node-2 /etc/init.d/Disabled-MonTier-es-raw-trans-Node-2
    mv /etc/init.d/MonTier-es-raw-trans-Node-3 /etc/init.d/Disabled-MonTier-es-raw-trans-Node-3
    mv /etc/init.d/MonTier-es-raw-trans-Node-4 /etc/init.d/Disabled-MonTier-es-raw-trans-Node-4
    
    mv /etc/init.d/MonTier-Derby /etc/init.d/Disabled-MonTier-Derby
    
    mv /etc/init.d/MonTier-HK-ESRetention /etc/init.d/Disabled-MonTier-HK-ESRetention
    
    mv /etc/init.d/MonTier-HK-SyslogKeepalive /etc/init.d/Disabled-MonTier-HK-SyslogKeepalive
    mv /etc/init.d/MonTier-HK-WsmKeepalive /etc/init.d/Disabled-MonTier-HK-WsmKeepalive
    
    mv /etc/init.d/MonTier-HK-WdpDeviceResources /etc/init.d/Disabled-MonTier-HK-WdpDeviceResources
    mv /etc/init.d/MonTier-HK-WdpServiceResources /etc/init.d/Disabled-MonTier-HK-WdpServiceResources
    
    mv /etc/init.d/MonTier-Reports /etc/init.d/Disabled-MonTier-Reports
    
    mv /etc/init.d/MonTier-UI /etc/init.d/Disabled-MonTier-UI
    
    sed -i -e "s/^SERVICES_FIRST_GROUP=\".*/SERVICES_FIRST_GROUP=\"\"/g" /etc/sysconfig/MonTier
    sed -i -e "s/^SERVICES_SECOND_GROUP=\".*/SERVICES_SECOND_GROUP=\"\"/g" /etc/sysconfig/MonTier
    sed -i -e "s/^SERVICES_THIRD_GROUP=\".*/SERVICES_THIRD_GROUP=\"\"/g" /etc/sysconfig/MonTier
    sed -i -e "s/\MonTier-HK-WdpServiceResources MonTier-HK-WdpDeviceResources//g" /etc/sysconfig/MonTier
    sed -i -e "s/^SERVICES_SEVENTH_GROUP=\".*/SERVICES_SEVENTH_GROUP=\"\"/g" /etc/sysconfig/MonTier
    Note: some errors might appear for services that are not exists in your specific deployment architecture type - for example "mv: cannot stat ‘/etc/init.d/MonTier-es-raw-trans-Node-4’: No such file or directory"
    In the Remote Collector
    Using any text editor (like vi), edit /etc/hosts files (e.g. vi /etc/hosts)
    Change the following entries:
    montier-es from 127.0.0.1 to the ip of the Local Node device
    Image Removed
    In the Remote Collector
    Using the Command Line Interface choose option 1 - "Start All", and wait until all the services are stopped, this may take a few minutes to complete.

    Verify in the console in Management → Internal health → Agents that all agents are in green state.
  2. Run the following two scripts, you will need to obtain them from IBM support:
    in the Local Node - configure_local_node.sh
    in the Remote Collector - configure_remote_collector.sh
  3. In the Local Node - !! Only if DPOD was already attached to DataPower Gateways !!
    you will need to reconfigure again all the the attached device.

...