\\ **[[cluster:0|Back]]** Note #1 CentOS 8.1 with the standard firewalld.\\ If this is of interest to you this was how I managed to get it work: EXTIFACE=MASTER_NODE_EXT_INTERFACE_DEVICE (e.g. eno1) INTIFACE=MASTER_NODE_INTERNAL_INTERFACE_DEVICE (e.g. eno2) INTIPADDR=MASTER_IP_OF_INTERNAL_IFAC PREFIX=PREFIX_OF_INTERNAL_NETWORK firewall-cmd --change-interface=${EXTIFACE} --zone=public firewall-cmd --change-interface=${INTIFACE} --zone=trusted --permanent firewall-cmd --permanent --direct --passthrough ipv4 -t nat -I POSTROUTING -o ${EXTIFACE} -j MASQUERADE -s ${INTIPADDR}/${PREFIX} firewall-cmd --set-default-zone=trusted firewall-cmd --reload And make sure the default route is set on all compute nodes. Note #2 configured Shorewall on a cluster to do NAT through the head node. Edit the file /etc/shorewall/snat and add this line: MASQUERADE 192.168.0.0/24 eno1 where 192.168.0 is the address range of your node interfaces - clearly you need to change this to fit en01 is the external interface on the head node My /etc/shorewall/interfaces contains this (forwarding ib0) nat eno1 detect dhcp nat ib0 detect dhcp so substitute ib0 for your internal ethernet interface ==== NAT Story, part 2 ==== This is my second NAT story, for the first one look at [[cluster:51|The Story Of NAT, part 1]] Writing this up so I will remember what I did, and why. Basic problem is this: How do you make a filesystem in a public VLAN available on a private network? One solution is to work with Network Address Translation, or NAT in short. More information at [[http://en.wikipedia.org/wiki/Network_address_translation|http://en.wikipedia.org/wiki/Network_address_translation]] We have a storage device which we refer to as flexstorage.wesleyan.edu which serves up a file system on login node petaltail. [root@petaltail ~]# host flexstorage flexstorage.wesleyan.edu has address 129.133.24.81 [root@petaltail ~]# df -h /home/dlbgroup Filesystem Size Used Avail Use% Mounted on flexstorage.wesleyan.edu:/share/dlbgroup 1000G 588G 413G 59% /home/dlbgroup Host petaltail has the following interfaces. The file system in question is mounted on host petaltail as VLAN 1 can reach VLAN 24. eth0 Link encap:Ethernet HWaddr 00:18:8B:51:FA:42 inet addr:192.168.1.217 Bcast:192.168.255.255 Mask:255.255.0.0 eth1 Link encap:Ethernet HWaddr 00:18:8B:51:FA:44 inet addr:10.10.100.217 Bcast:10.10.255.255 Mask:255.255.0.0 eth2 Link encap:Ethernet HWaddr 00:15:17:80:8D:F2 inet addr:129.133.1.225 Bcast:129.133.1.255 Mask:255.255.255.0 eth3 Link encap:Ethernet HWaddr 00:15:17:80:8D:F3 inet addr:192.168.2.2 Bcast:192.168.2.255 Mask:255.255.255.0 But a compute node on our cluster, for example node b1, has the following interfaces, all private eth0 Link encap:Ethernet HWaddr 00:13:D3:F2:C8:EC inet addr:192.168.1.7 Bcast:192.168.255.255 Mask:255.255.0.0 eth1 Link encap:Ethernet HWaddr 00:13:D3:F2:C8:ED inet addr:10.10.100.7 Bcast:10.10.255.255 Mask:255.255.0.0 So in order to for the compute node b1 to reach the flexstorage server we need to use NAT rules and define a path/route. First we start on petaltail and edit the iptables file and add a "nat filter" masquerade/post routing directives and in the "filter filter" set up a rule connecting eth1 and eth2. *nat # fss public to 10.10 -A POSTROUTING -o eth2 -j MASQUERADE COMMIT *filter # fss public via 10.10 -A FORWARD -i eth1 -o eth2 -m state --state RELATED,ESTABLISHED -j ACCEPT ... COMMIT Next, on the compute nodes we need to add routing path and then mount the file system (using an IP because there is no name resolving). These commands are stuck in /etc/rc.local for persistence. # /etc/rc.local route add -host 129.133.24.81 gw 10.10.100.217 eth1 mount 129.133.24.81:/share/dlbgroup /home/dlbgroup -t nfs -o soft,intr,bg [root@b1 ~]# df -h /home/dlbgroup Filesystem Size Used Avail Use% Mounted on 129.133.24.81:/share/dlbgroup 1000G 588G 413G 59% /home/dlbgroup There is ofcourse a penalty in performance doing this. [root@petaltail ~]# time dd if=/dev/zero of=/home/dlbgroup/foo bs=1024 count=1000000 1000000+0 records in 1000000+0 records out 1024000000 bytes (1.0 GB) copied, 107.961 seconds, 9.5 MB/s real 1m47.964s user 0m0.322s sys 0m2.094s [root@b1 ~]# time dd if=/dev/zero of=/home/dlbgroup/foo bs=1024 count=1000000 1000000+0 records in 1000000+0 records out 1024000000 bytes (1.0 GB) copied, 110.017 seconds, 9.3 MB/s real 1m50.027s user 0m0.271s sys 0m4.073s \\ **[[cluster:0|Back]]**