r10 - 13 Aug 2011 - 11:30:27 - SaulYoussefYou are here: TWiki >  Admins Web > SupportingGLOW

SupportingGLOW

GLOW VO

  • The Grid Laboratories of Wisconsin, a VO for the University of Wisconsin's Center for High Throughput Computing: http://chtc.cs.wisc.edu/
  • Contact: Steve Barnet <barnet@cs.wisc.edu>, Brooklin Gore <BGore@Morgridgeinstitute.org>, Dan Bradley <dan@hep.wisc.edu>, chtc@cs.wisc.edu
  • Submission system: GlideinWMS

Requirements

  • Worker node OS should be RHEL 5 (or equivalent, such as SL5)
  • Worker nodes require outbound internet access (nodes can be behind NAT)
  • Worker node memory: 1GB / slot minimum, assume 1.5-2 GB
  • Worker node scratch: assume 10G per job slot - though typically using less than 1GB
  • The worker nodes require the OSG CA certs that are installed as part of the OSG Worker Node Client. Host certs on the worker nodes are not required.
  • Site squid (optional): used for data in some jobs ~200 MB
  • Job preemption is okay
  • glexec is not required

Workflow and Security profile

Description:

  • For every job slot, a pilot job process starts up.
  • The pilot job sends outbound TCP traffic over HTTP to a host at UCSD ("factory") and via HTTP to a host at wisc.edu ("frontend").
  • The pilot job spawns a condor startd, which spawns a condor starter.
  • The startd sends outbound UDP traffic to a single port on the frontend. This is randomly chosen (per pilot) from a range of 200 ports currently. This can be changed to TCP if necessary.
  • The startd and starter send outbound TCP traffic to two ports on one of the Condor submit machines at wisc.edu, to communicate with the condor_schedd and condor_shadow. The ports are chosen randomly (per glidein) from the host's ephemeral ports. The choice of ports can (and in some cases has) been reduced to a single static port, which is typically 9618.

Hosts and ports:

  • The frontend is glidein.chtc.wisc.edu (128.104.59.136)
    • ports: 80, 9618-9820
  • The factory is glidein-1.t2.ucsd.edu
    • port 8319
  • The Condor submit machines (more may be added in future)
    • submit.chtc.wisc.edu
    • login01.hep.wisc.edu
    • poplar.hep.wisc.edu
    • spaldingsubmit.botany.wisc.edu
    • simon.stat.wisc.edu
    • chtcsubmit.ansci.wisc.edu
    • condor.morgridge.net

Setup

Status

GLOW VO status


Notation: led-green completed led-blue work is in progress led-red table to be updated

Site OSG Resource Status Job Max
TIER1 gridgk01.racf.bnl.gov led-green 500
AGLT2 gate02.grid.umich.edu led-green 50
MWT2 osg-gk.mwt2.org led-red 125
NET2_BU   led-green 15
UTA_SWT2   led-red 10
SWT2_OU_OSCER grid1.oscer.ou.edu/jobmanager-lsf led-green 64
WT2   led-red  

Known Issues


-- RobertGardner - 02 Aug 2011

About This Site

Please note that this site is a content mirror of the BNL US ATLAS TWiki. To edit the content of this page, click the Edit this page button at the top of the page and log in with your US ATLAS computing account name and password.


Attachments

 
Powered by TWiki
This site is powered by the TWiki collaboration platformCopyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki? Send feedback