glideinWMS. This exercises a lot of HTCondor features that are important to OSG-type users: glidein, HTCondor-G, CE schedd, CCB. Igor suggested that it would be a good idea of we could run this test in-house, and we agree, so we have set up the machinery to do so. This document describes roughly how we installed it.
- ghost-pool - runs light-weight jobs on cluster nodes. The setup is in /unsup/condor/ghost-pool. Some nodes in CHTC also join the ghost pool. These are configured in the CHTC cfengine.
- CE: vdt-itb.cs.wisc.edu - an OSG CE that is part of the ITB (integration test bed). This node will be referred to as the CE. The schedd on this node is part of the ghost pool. The jobmanager is vdt-itb.cs.wisc.edu/jobmanager-nfslitecondor.
- factory: c157.cs.wisc.edu - the glideinWMS factory, installed in /data1/itb-stress. It submits glideins to the CE. This node is part of the GLOW pool (despite its domain name). It normally runs HTCondor jobs. Turn off HTCondor when doing a large-scale test.
- submit: c158.cs.wisc.edu - the glideinWMS VO frontend, user job submission node, and collector tree for the glidein pool, all installed in /data1/itb-stress. This node normally runs HTCondor jobs. Turn off HTCondor when doing a large-scale test.
glidein factory setup
Go to c157.chtc.wisc.edu. Install the factory in /data1/itb-stress. Where an unprivileged user account is needed, we are using zmiller.
Documentation for the glideinWMS factory may be found here: http://www.uscms.org/SoftwareComputing/Grid/WMS/glideinWMS/doc.v2/install/factory_install.html
submit node setup
The submit node is c158.chtc.wisc.edu.
The documentation for setting up the glideinWMS VO frontend and pool collector are here: http://www.uscms.org/SoftwareComputing/Grid/WMS/glideinWMS/doc.v2/install/
setting up CE
The CE on vdt-itb.cs.wisc.edu was already set up except for a HTCondor jobmanager.
SET UP HTCondor 1) installed condor-rhel5.repo into /etc/yum.repo.d 2) yum install condor 3) edit /etc/condor/condor_config? 4) edit /etc/condor/condor_config.local change CONDOR_HOST = chopin.cs.wisc.edu change COLLECTOR_HOST = $(CONDOR_HOST):15396 change DAEMON_LIST = MASTER, SCHEDD 5) export CONDOR_CONFIG=/etc/condor/condor_config 6) /usr/sbin/condor_master okay, that's working. SET UP JOBMANAGER 7) cd /opt/itb 8) DO NOT DO THIS: pacman -trust-all-caches -get http://software.grid.iu.edu/osg-1.1.14:Globus-Condor-Setup 9) do: [root@vdt-itb ~]# export CONDOR_CONFIG=/etc/condor/condor_config [root@vdt-itb ~]# export VDT_CONDOR_LOCATION=/usr [root@vdt-itb ~]# export VDT_CONDOR_CONFIG=/etc/condor/condor_config 10) cd /opt/itb/ 11) pacman -install http://vdt.cs.wisc.edu/vdt_2099_cache:Globus-CondorNFSLite-Setup 12) $VDT_LOCATION/globus/lib/perl/Globus/GRAM/JobManager/condornfslite.pm update condor_config location, condor_rm, condor_submit 13) edit $VDT_LOCATION/edg/etc/grid-mapfile-local, add DNs 14) run $VDT_LOCATION/edg/sbin/edg-mkgridmap