Production, SC4, and general ATLAS user analysis requirements - General requirements and schedules from: --> sc4 and csc schedules --> software/service schedules (panda and dq2 releases, other tools) --> other commissioning projects? - Validation of software and services on the Tier2 - Integration and validation with new OSG services - Specific requirements on the site: * for Panda jobs (eg, use of local SRM-based SE, web caching services, firewalls and ports, workernode requirements, local disk requirements, etc ). When does ws-gram come into play; implications for tier2. * for general ATLAS jobs -- guidelines and practices, publication to users * for OSG grid users * for local/interactive users Tier2 data services - model - which services are provided, do we have archetypes for deployment and operational support, interaction with physics working groups and analysis support centers (dataset provisioning and management, priorities) - ATLAS distribted databases (tag, conditions, geo, etc) & catalogs - web caching services - DDM related services - user access and usage - interaction of all the above with backend storage systems (dcache, srm, filesystems, etc, see below) Storage and DDM requirements for site storage, management - SRM/dCache production-level services * this is probably the most important issue * need to get on SAME footing with BNL Tier1 w.r.t. release, validation, availability of dccp clients, etc. - other local storage, home servers for users Site architecture and configuration --- review where we are, discussed modifications for next procurements -- perhaps this can come out in site reports. - CPU farms - Storage - Grid services - Edge servers - Interactive hosts - Hosts for Tier3 integration? - Isolating ATLAS and OSG workloads Monitoring and Accounting - what are precise monitoring requirements: which site-level systems, reporting to which top-level servers, etc. - service alerts and monitors - for operations and capturing and tracking ATLAS specific issues - accounting for RAC purposes - accounting for LCG purposes - accounting for OSG purposes Policy publication and implementation - GUMS configurations - Not only the queues, priorities, and quotas, but also Tier2 usage by general ATLAS users (analysis, production, interactive) - local scheduler configurations (PBS and Condor) - publication of US ATLAS and site-level policies. - policy change control OSG integration - running OSG services on Tier2 resources - integration / validation of ATLAS on OSG - supporting other VOs on usatlas resources - interfacing to OSG operations, security, etc LCG interoperability - how do we support this, at what level, and how do we account for it correctly? - providing usatlas bdii for this purpose, plus condor ClassAd generator Operations - User support: a basic plan for servicing local issues, interfacing with overall ATLAS operations groups - Coordination with, support for, analysis support centers - Admin support * Maintenance and upgrades in the presence of on-going SC's and production running. - DDM operations monitoring - Panda operations monitoring Network upgrades - follow up on upgrades or new issues from the networking workshop - combined with discussion of storage services, plan data I/O tests and identify bottlenecks and plans to address - network tuning parameters - recommendations Tier3 model - can we flesh out a bit, even crudely - how should tier2's support these facilities. - what is the model for data access, code dev, in relation to datasets hosted at tier2, eg. - other tier2 support for tier3 facilties-- RobertGardner - 26 Apr 2006
Please note that this site is a content mirror of the BNL US ATLAS TWiki. To edit the content of this page, click the Edit this page button at the top of the page and log in with your US ATLAS computing account name and password.