...
- A projectbook like we did for USNO could be appropriate
- Process diagrams (how systems boot, how jobs get started from NRAO and run, how remote institutions start jobs, etc)
Networking
HTCondor flocking requires
- From local schedd to remote collectord on condor port 9618
- From remote negotiator and execute hosts to local schedd. Here the execute hosts can be NATed.
- From local shadow to remote starterd. Use CCB. It allows execute hosts to live behind firewall and be NATed.
NRAO side
- NRAO -> remote head node on port 22 (ssh)
- Submit Host -> remote head node (condor_collector) on port 9618 (HTCondor) for flocking
- Submit Host <- remote head node (condor_negotiator) on port 9618 (HTCondor) for flocking
- mcilroy has external IPs (146.88.1.66 for 1Gb/s and 146.88.10.66 for 10Gb/s). Is the container listening?
- Submit Host <- remote execute hosts (condor_starter) on port 9618 (HTCondor) for flocking
- Submit Host (condor_shadow) -> remote execute hosts (condor_starter) on port 9618 (HTCondor) for flocking. CCB might alleviate this.
Remote side
- Head node <- from nrao.edu on port 22 (ssh)
- Head node -> revere.aoc.nrao.edu on port 25 (smtp)
- Head node -> NRAO Submit Host on port 9618 (HTCondor) for flocking
- Head node <- NRAO Submit Host on port 9618 (HTCondor) for flocking
- Execute node -> NRAO Submit Host on port 9618 (HTCondor) for flocking. Execute host may be NATed.
- Execute node -> gibson.aoc.nrao.edu on port 22 (ssh) for flocking with nraorsync. Execute host can be NATed.
...