The available disk on the MD1200 boxes is allocated in 3 equal parts (dCache, Hadoop, xrootd).
The R610 will use one drive to run the host OS (SL 5.5) and the other 5 drives (RAID configuration - LVS) will be used for the virtual machines. There are 9 VMs, 3 each for dCache, Hadoop, xrootd
Storage (virtual) nodes include:
3 VMs for dcache pools
3 VMs for xrootd pools
3 VMs for hdfs pools
New machines 2 (8 R610)
These R610 servers will be installed with:
VMware ESXi on itb-esx3 (installation CD image available from vmware.com)
SL 5.5 and KVM in itb-kvmX: one drive to run the host OS (SL 5.5) and the other 5 drives (RAID configuration - LVS) will be used for the virtual machines
SL 5.5 like Tier 2/3 worker nodes on itb-cXXX: probably one drive to run the OS (SL 5.5) and the other 5 drives (RAID configuration - LVS) for the scratch space
for itb-dyn check what is needed for dynamic provisioning (create/destroy VM dynamically); check Nimbus requirements (http://www.nimbusproject.org/ )
Headnode services for ITB and the virtual tier 3 include:
PBS scheduler
Condor headnode (collector/negotiator)
GUMS
ITB CE Headnode
VTB CE Headnode
Tier 3 CE
Tier 3 SE (server varies)
Tier 3 GUMS
Tier 3 UI
VO frontend (for Glidein-WMS job submission)
Engagement submit host (with OSG MM submission)
Campus router
We have currently 12 services in the list above. We may consider allocating one of the 2 unallocated R610 to host headnode services as well (itb-kvm6.uchicago.edu, with int/ext cabling).
We discussed having one management/service node (possibly one of the previous ITB nodes) to use as:
login node
NFSv4 server for home directories
Ganglia monitor
Services "borrowed" from the Tier 2/Tier 3 infrastructure:
Cobbler
OSG_APP with ATLAS releases
IP Reallocation
IPs from the grid school machines have been reallocated as follows:
Testing
Interesting tests that we discussed.
IO testing
IO comparison:
local IO (on host OS) vs
direct IO from VM (aka raw disks, access to the HW partition) vs