9f46457324d46c3b72937c4d52879f55.ppt
- Количество слайдов: 18
Site report: CERN Helge. Meinhard (at) cern. ch HEPi. X/HEPNT spring 2004 Edinburgh
Structure, management n n As of 01 January 2004: DG (Aymar), CSO (Engelen), CFO (Naudi), 7 Departments IT Division has become IT Department n n n Head: W. von Rüden, deputy: J. Ferguson All of previous IT Division Major part of AS Division: 3 groups working on AIS n n n 2 groups merged (R. Martens) 1 group merged with DB group Printshop, document support from ETT Division n Merged with user support (M. Draper) 2
Computer Security n AFS password expiration in force n n n 12 months, like other services (e. g. NICE) Insecure protocols (e. g. ftp) banned for offsite traffic Biggest challenge are viruses and worms n Visitor / unmanaged PCs, dual-boot systems 3
Fabric Infrastructure and Operations (1) n Quattor / Lemon deployment n n LEAF n n Moving ahead, major benefits German Cancio’s talk Lemon status display, combined metrics Miroslaw Siket’s talk State Management System taking shape, being integrated with other tools System administration n 2 nd level of 3 -level support model Team now 7 people In full swing, all major clusters under their responsibility 4
Fabric Infrastructure and Operations (2) n CC refurbishment ongoing n n Civil engineering work for bunker for new 18 k. V substation completed Right-hand side of large machine room ready Emptying left-hand side has begun, to be completed by July Disk and tape servers: re-installed with standard configuration n More details: Tim Smith’s talk 5
Fabric Infrastructure and Operations (3) n Procurements n n Serial consoles n n Tenders open for white boxes (presumably…) and SATA-based disk servers Preparing tenders for disk arrays for autumn Documents for Market Survey for purchases in 2006… 2008 drafted All hardware procured (LSZH cables - major amount of work) Decision to use SLAC software, collaboration with Chuck First machines wired up, first users relying on service Stress tests for new and repaired machines n n n Following implementation at SLAC Now part of standard procedure Handled by Sysadmin team 6
Architecture and Data Challenges (1) n CERN Openlab n n n Workshops on total cost of ownership, and on security held Oracle joined as full partner, Voltaire joined as a contributor Detailed studies n n Fast interconnects Disk server performance: Choice of RAID, file system, . . . Jan Iven’s talk Storage performance measurements at Caspur 7
Architecture and Data Challenges (2) n Linux certification: CEL 3 nearing completion n n n CERN recompile of RHEL 3 sources Aim to provide lxplus/lxbatch like service by end May 2004 CERN 7. 3 will be supported until end 2004 Jarek Polok’s talk Considering (and testing) RHEL Panel and discussion Data challenges of CMS and Alice completed, Atlas and LHCb started New stager, Storage Resource Manager Olof Bärring’s talk 8
Data Bases (1) n n To date: too much depends on good will and good luck Vision: Taking advantage of reorganisation n n n Re-evaluate services Streamline architectures, configurations, processes Simplify management, maintenance and troubleshooting Improve security, test regularly SLAs (realistic, measurable) and quality control Required for scalability in the future 9
Data Bases (2) n Openlab work: 2 Oracle-funded fellows n n Evaluation of Oracle 10 g Database Application Server and Enterprise Manager Features to evaluate: Cross-platform transportable spaces n Stand-by data bases n Replications/streams n Native numbers n EZ install n 10
Data Bases (3) n LCG services – data challenges n CMS DC 04 (April/May 2004) n n Smooth running on Physics Sun Cluster after close collaboration with users First full-scale production usage of LCG file catalog (RLS): performance problem of middleware, IT-GD working Other data challenges starting, continuing to stress the Physics Sun cluster and the RLS Future: investigate Oracle Streams for LCG file catalog replication, deploy stand-by DB and (later) Oracle application server and DB clusters 11
Data Bases (4) n POOL: Persistency framework for LHC n n Use Root I/O to stream data to files Successfully deployed during CMS DC 04, Atlas and LHCb to test in their DCs n n n Alice keeps using Root directly Workarounds implemented to cope with RLS middleware performance limitations New Oracle contract n n n Based on named users Platform and location independent More Application Server licences Maintenance costs reduced and fixed for 9 years Extended to all CERN staff + users (includes remote usage for CERN related work) Distribution has been prepared (support concerns) 12
Grid Deployment n n LCG status Oliver Keeble’s talk LCG user registration, VO management Maria Dimou’s talk 13
Internet Services n n n Mail migration (to Exchange servers) completed – 14000 users migrated Windows terminal servers adopted as a service, 227 users during first months Access to CERN DFS via Web. DAV adopted as a service Windows screen saver for production fieldproven, deployment imminent Fighting effects of Sasser worm 14
Product Support (1) n Contract for distributed computing support retendered n n n Won by SERCo Being prepared to be in full operation by 1 st July 2004 CVS services n n n Two flavours, one with repositories on AFS (more failsafe), one with local repositories – 4 machines each Major projects migrated to this service (e. g. Atlas offline) Manuel Guijarro’s talk 15
Product support (2) n Solaris n n Ximian connector n n Status and plans Manuel Guijarro’s talk Source code released by Novell Direct MAPI access, including advanced features such as shared calendaring PS evaluating deployment on Linux and Solaris using Evolution as mail client Busy preparing CHEP… 16
User and Document Services n n In. Di. Co: Web application for organising conferences Mick Draper’s talk Major changes to Computing Helpdesk as a consequence of retendered Distributed Computing Support contract n More people at centralised (building 513) helpdesk, less local support 17
Communication Systems n n Internet Land Speed record: 6. 25 Gb/s between Los Angeles and Geneva Smooth migration to new GSM operator (Sunrise) Portable registration enforced ACB (Automatic call-back) shutdown: n n n 01 -Jul-2004: Call-in only 31 -Dec-2004: End of service Recommended replacement: ISPs on open market 18


