You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 66 Next »

Operating System

  • Must support CASA
  • Will need a patching/updating mechanism
  • How to boot diskless OS images
    • I am not finding any new sexy software packages to automate PXE+DHCP+TFTP+NFS
    • One OS image for both our use and locals use or multiple OS images?
    • Use containers (docker, singularity, kubernetes, mesos, etc)?
    • Ask Greg at CHTC what they use
  • What Linux distrobution to use?
    • Can we use Red Hat with our current license?
    • Should we buy Red Hat licenses like we did for USNO?
    • Do we even want to use Red Hat?
    • Rocky Linux or AlmaLinux since CentOS is essentially dead?

Third party software for VLASS

  • CASA
  • HTCondor
  • Will need a way to maintain the software

Third party software for Local

  • Will need a way to maintain software for the local site

Services

Management Access

  • PDU
  • UPS
  • BMC/IPMI
  • switch

Maintenance

  • replace disk (local admin)
  • replace/reseat DIMM (local admin)
  • replace power supply (local admin)
  • NRAO may handle replacement hardware. Drop ship. Spare ourselves?
  • Patching OS images (NRAO)
  • Patching third party software like CASA and HTCondor (NRAO)
  • Altering OS images (NRAO)

Hardware


Using

  • Get NRAO jobs on the remote racks.  This may depend on how we want to use these remote racks. If we want them to do specific types of jobs then ClassAd options may be the solution. If we want them as overflow for jobs run at NRAO then flocking may be the solution. Perhaps we want both flocking and ClassAd options.  Actually flocking may be the best method because I think it doesn't require the execute nodes to have external network access.
    • Flocking?
    • Classad options?
    • Other?
  • Remote HTCondor concerns

    • Do we want our jobs to run a an NRAO user like vlapipe or nobody?
    • Do we want local jobs to run as the local user, some dedicated user, or nobody?


  • Need to support 50% workload for NRAP and 50% workload for local.  How?
  • Share disk space on head node 50% NRAO and 50% local
    • Two volumes: one for NRAO and one for local?

Documentation

  • A projectbook like we did for USNO could be appropriate
  • Process diagrams (how systems boot, how jobs get started from NRAO and run, how locals start jobs, etc)


Other

  • Keep each rack as similar to the other racks as possible.
  • Test system at NRAO should be one of everything.

Since we are making our own little OSG, should we try to leverage OSG for this or not?  Or do we want to make each POD a pool and flock?

Should we try to buy as much as we can from one vendor like Dell to simplify things?

APC sells a packaged rack on a pallet ready for shipping.  We could fill this with gear and ship it.  Not sure if that is a good idea or not.



Resources

  • USNO correlator (Mark Wainright)
  • VLBA Control Computers (William Colburn)
  • Red Hat maintenance (William Colburn)
  • Virtual kickstart (William Colburn)
  • HTCondor best practices (Greg Thain)
  • OSG (Lauren Michael)


Site Questions

  • Door width and height and path to server room.  Can a rack-on-pallet fit?  Can it fit upright on casters?
    • NRAO-NM wide server door is 48"W x 84"H
  • Voltage in server room (110V or 208V or 240V)
  • Receptacles in server room (L5-30R or L21-30R or ...)
  • Single or dual power feeds
  • Firewalls




  • No labels