Скачать презентацию The INFN Regional Center s as main node s of Скачать презентацию The INFN Regional Center s as main node s of

8d7ee85ca6e9041ef835ecb63431ba93.ppt

  • Количество слайдов: 33

The INFN Regional Center(s) as main node(s) of the INFN Grid Mirco Mazzucato On The INFN Regional Center(s) as main node(s) of the INFN Grid Mirco Mazzucato On behalf of Expts Comp. Coord. and tech experts(~EB of INFN-Grid) Paolo Capiluppi, Domenico Galli, Alberto Masoni, Laura Perini, Fulvio Ricci, Antonia Ghiselli, Federico Ruggieri 4/12/00 M. Mazzucato – GR 1 - Roma 1

Main conclusion of the “Hoffmann Review” q The Panel 1 recommends the multi-tier hierarchical Main conclusion of the “Hoffmann Review” q The Panel 1 recommends the multi-tier hierarchical model proposed by Monarc as one key element of the LHC computing model with the majority of the resources not based at CERN : 1/3 in 2/3 out q About equal share between Tier 0 at CERN, Tier 1’s and lower level Tiers down to desktops q Tier 0 / S(Tier 1) / S (all Tier 2 +… ) = 1 /1 /1 q General consensus that GRID technologies developed by Datagrid can provide the way to efficiently realize this infrastructure q All experiments should perform Data Challenges of increasing size and complexity until LHC start-up involving also Tier 2 q EU Testbed : 30 -50% of one LHC experiment by 2003 (Match well with INFN Grid assumptions : 10% of final size for each experiment) q Limit heterogeneity : OS = Linux + backup solution, Persistency = 2 tools max 4/12/00 M. Mazzucato – GR 1 - Roma 2

HEP Regional Centre Hierarchy CERN Tier 0 Tier 1 France Italy 2. 5 Gbps HEP Regional Centre Hierarchy CERN Tier 0 Tier 1 France Italy 2. 5 Gbps UK 2. 5 Gbps Fermilab ~Gbps Tier 2 center Tier 2 Tier 3 622 Mbps Site 100 Mbps-1 Gbps Tier 4 4/12/00 desktop M. Mazzucato – GR 1 - Roma INFN-GRID 3

Tier Computer Centers functionalities (LHC Computing Review) The Monarc model (first approximation) definitions: q Tier Computer Centers functionalities (LHC Computing Review) The Monarc model (first approximation) definitions: q Tier 0(Cern): raw data storage and first calibration + reconstruction. Very large storage capacity (MSS) q Tier 0+Tier 1 at CERN : 10 PB/yr tapes; 2 PB/yr disk; 2 M SI 95 q Tier 1: further calibration+reconstruction passes. Large fraction of simulation and analysis. Large storage capacitiy. Associated support. All services. q Typical Tier 1 : 3 PB/yr tapes; 0. 5 PB/yr disk, 0. 9 M SI 95 q Tier 2 and lower levels: simulation, analysis. q Typical Tier 2 = 20 -30% of Tier 1 q Tier 0 and Tier 1 are basically open to all members of a Collaboration, a Mo. U will specify the conditions, Tier 1 will then be needed for the lifetime of LHC q Detailed quantitative estimates of resources and costs are not touched in this talk, but see F. Ruggieri talk for preliminary INFN estimates 4/12/00 M. Mazzucato – GR 1 - Roma 4

R&D in Italy : THE INFN-GRID PROJECT q The proposal has been submitted to R&D in Italy : THE INFN-GRID PROJECT q The proposal has been submitted to INFN management and CSNs at the end of July and presented at CSN in September q The size of the project : 26 Sites, ~200 people, ~ 70 FTE’s q The requested funding for 3 years : ~ 10 M Euro q Activities are organized at national level around WP’s developed by Datagrid (9. 8 ME), testbeds and experiments applications. Aim to provide multi-tier integration. q Large scale testbeds provided by LHC experiments (Regional Center prototypes) + Virgo q Hope to receive recommendation of approval by CSNs <2000 q Preliminary reserved financing of basic HW infrastructure and R&D: ~ 2 M Euro’s for year 2001 : OK compared to EU partners 4/12/00 M. Mazzucato – GR 1 - Roma 5

Strategy of INFN-Grid q Foster close collaboration between LHC experiments and computer professionals (~25 Strategy of INFN-Grid q Foster close collaboration between LHC experiments and computer professionals (~25 FTE’s in INFN Grid) Software and Middleware q Evaluation and deployment of existing GRID basic services (Globus, Condor, …) mainly looking for production quality q Robustness, scalability (hundreds of users, hundreds of jobs to run, huge data sets, …), reliability very important requirements q Use of GRID software for real applications and on real experiment production environments (CMS already started, Alice, ATLAS, LHCb coming soon. . ) q Implement missing functionalities in close collaboration with US teams q Incremental deliveries q Tight collaboration with EU partners in Datagrid to q Develop HEP middleware to implement multi-Tier Monarc model for expts q Coordinate relations with US projects: Globus, Gri. Phy. N, PPDG… 4/12/00 M. Mazzucato – GR 1 - Roma 6

Strategy of INFN-Grid (Cont. ) Hardware infrastructure q Deploy testbeds using resources of “all” Strategy of INFN-Grid (Cont. ) Hardware infrastructure q Deploy testbeds using resources of “all” INFN sites, connected to Data. Grid testbed as recommended by Panel 1 of “Hoffmann” Review : ”. . involve most of the final distributed computing system” q Develop Computing Fabric prototypes to understand issues related to : q Architecture q Hardware choices q Performance Vs. cost of different disk technologies : SAN/RAID/SCSI/EIDE = 10/5/3/1 in cost q Can we do everything with EIDE disks ? Which tech. is more adequate for each Data set ? q Set up, maintenance and support of large systems q Define network topology and evaluate net technology services 4/12/00 M. Mazzucato – GR 1 - Roma 7

The target prototypes Regional Centers for LHC experiments in INFN-Grid q All 4 LHC The target prototypes Regional Centers for LHC experiments in INFN-Grid q All 4 LHC experiments have chosen to concentrate in this prototyping phase (20012003) the bulk of computing resources in one place q ATLAS : Rome q CMS : Legnaro INFN National Laboratories q Alice : Torino q LHC-b(2001) : Bologna q But since INFN has a positive advanced experience of distributed computing (e. g. Condor pool INFN) and since INFN manpower (Physicists and Computing experts) is spread out, several Tier 1/2 functionalities are supported by few different sites : q Atlas : Milano q CMS : Bologna, Bari, Padova, Pisa and Rome q Alice : Bologna, Bari, Catania q No common discussions were done in INFN Grid about a possible solution for the final Tier 1 centers. Decision was postponed at the end of the prototyping phase. q Note: “Hofmann” Panels conclusions Comp. Mo. U end 2001 INFN management and CSN: anticipate decision on RC’s 4/12/00 M. Mazzucato – GR 1 - Roma 8

 “Hoffmann” Review outcome relevant for INFN RC’s q For the first time planned “Hoffmann” Review outcome relevant for INFN RC’s q For the first time planned an integrated world-wide computing system : q Tier 1 s open to all members of the collaboration (Up time and efficiency!!) q Very large weight given to Tier(>=2) Centers (Contrary to the past!) q Share of resources between Tier 1 and Tier(>2) centers is recognized to be an internal national affair q CERN IT will provide and support, as for LEP, only the basic SW and HW infrastructure at CERN: Tier 0 +Tier 1 for each experiment. “Budget limited” personnel is foreseen to support experiments issues concerning the usage of this infrastructure. q As for LEP apparently there will be no IT support on specific experiment’s activities : software distributions, productions, simulations, data replications and distribution, analysis etc. These are considered internal experiments issues 4/12/00 M. Mazzucato – GR 1 - Roma 9

The issue: One vs more Tier 1 in INFN Some preliminary considerations…. q Technology The issue: One vs more Tier 1 in INFN Some preliminary considerations…. q Technology (almost all commodity and scalable) q Computing and storage fabric are built up from commodity components q Simple PC’s q Inexpensive network attached disks q Standard network interfaces and probably Standard Lan backbone (Whatever Fast-Giga Ethernet will be in 2006) q High bandwidth WAN connections may not become a commodity but EU Community and Countries will probably support high bandwidth Research Network for strategic reasons (see EU Geant project, Garr-B…. ) q Mass Storage will probably not be a commodity. . but future Optical Storage Very easy to split an re-group according to needs WAN in 2006 at Gbits like present PC bus Different from 1988 when the scene was dominated by mainframes Assume MSS is the only large not scalable single piece remaining Do we need it in INFN? Need technical evaluation ! 4/12/00 M. Mazzucato – GR 1 - Roma 10

One vs more Tier 1 in INFN (Cont. ) Can we learn from other One vs more Tier 1 in INFN (Cont. ) Can we learn from other experiences ? q A lot from experienced HEP computing centers (CERN. . , Babar, CDF… ) q Technology, manpower…. . q But need to adapt computing model to INFN conditions q INFN availability of computing professionals and know how is not negligible at all (~25 FTE’s in Grid) but spread out in many different sites (26 in INFN Grid) q INFN Computing Services are used to support computing facilities and network connections (A lot of people and expertise…but again distributed) q INFN-Grid provides a framework for a large common cooperative effort in computing. Common project allows to best profit of all distributed expertise. q Very positive up to know. Sharing of work between sites seems as effective as for detector construction. q In the past very successful experience of collaboration INFN wide was done with INFNet, Garr 2 and Garr-B planning, INFN Condor pool. q Very important role again played by CNAF in providing central coordination and support to Datagrid middleware development and EU+INFN testbed deployment. 4/12/00 M. Mazzucato – GR 1 - Roma 11

One vs more Tier 1 in INFN (Cont. ) q INFN is unique in One vs more Tier 1 in INFN (Cont. ) q INFN is unique in Europe (more similar to US) q. In France expertise and manpower mostly concentrated in Lyon q. Build up in 80’s around a non scalable mainframe q. . and very limited WAN badwidth q. Try now to establish more sharing (Marseille, Grenoble. . ) q. In UK computing professional manpower today is very limited. q. In Germany not a central organization like INFN (Computing mainly done in University Computing Centers…like CINECA or CILEA) and DESY is not involved q. In Netherland computing skill mostly concentrated in SARA (Multi-discipline CC). . 4/12/00 M. Mazzucato – GR 1 - Roma 12

OUR CONCLUSION ON RC’s q The choice should be made with the only and OUR CONCLUSION ON RC’s q The choice should be made with the only and unique objective to provide the most efficient and competitive way to perform data analysis for INFN groups taking into account the INFN reality and exploiting at best all available resources. q Watch at other experiences, but no blind duplication!! q The choice should take into account as much as possible q INFN structure q Up time and efficiency q Experiment computing model q Technology evolution and Costs q The solution should provide a complete detailed plan of implementation taking into account role and commitments of the different sites, manpower available and foreseen, expertise etc. 4/12/00 M. Mazzucato – GR 1 - Roma 13

OPTIONS TO BE EVALUTED In the first meeting it was decided to limit the OPTIONS TO BE EVALUTED In the first meeting it was decided to limit the evaluation to 3 options q 1 TIER-1 MULTI-EXPERIMENT (Alice, Atlas, Cms, Lhc-B, Virgo) q Located in Bologna area in a new (built or rent) computing facility q Set up by CNAF q Preliminary document prepared by Federico Ruggieri (yet to be discussed) q 2 TIER-1 MULTI-EXPERIMENT q Located in National labs of Frascati and Legnaro q Documents in preparation q 3 (or more) TIER-1 according to experiment whishes q Compared to hyp 2 add (at least) 1 Tier 1 in Torino General consensus that Tier 1 have to be considered as special Grid node(s) which will provide to other experiment nodes (Tier 2. . Tier 4) the functionalities that are missing or not convenient to duplicate (e. g. mass storage, system support manpower etc. ). What is important is the integrated throughput of the overall system and not the one of one single component. 4/12/00 M. Mazzucato – GR 1 - Roma 14

Preliminary set of questions For each of the 3 alternatives expts were asked to Preliminary set of questions For each of the 3 alternatives expts were asked to provide a document answering the following preliminary set of questions q Global estimate of resources needed in Italy / Total expt. q Provide evaluation and detailed computing model containing q Role, resources , manpower etc. of each candidate Tier(n) site q Relations between Tier-1 e Tier-n (functionalities and localization) q q q q Need of mass storage and localization Space and services available and to be acquired Available manpower by expts and Computing Services to support Tier Manpower to support experiment applications (Sim, Rec, An. ) Implications on network connectivity of Tier location Manpower for expts user support Cost estimation Candidates Tier 1 sites asked to produce a document containing estimation of space, available and total manpower, costs etc. 4/12/00 M. Mazzucato – GR 1 - Roma 15

Present Status q Too short time for expts and candidates centers to produce documents Present Status q Too short time for expts and candidates centers to produce documents discussed and agreed within the collaborations q A lot of activities going on: q Hoffman Review q Datagrid q INFN Grid q Regional Centers prototypes q Need to define timescale to have study and proposal(s) completed (CMS: June 1 rst ? ) q Very preliminary answers received : raw material coming independently from each expts. Not at all discussed yet ! q Shown to give flavor of starting views. 4/12/00 M. Mazzucato – GR 1 - Roma 16

MODELLI CONSIDERATI ITALIA q LA COLLABORAZIONE ALICE HA DISCUSSO (21 -22/02/00) LE POSSIBILI ALTERNATIVE MODELLI CONSIDERATI ITALIA q LA COLLABORAZIONE ALICE HA DISCUSSO (21 -22/02/00) LE POSSIBILI ALTERNATIVE PER I CENTRI REGIONALI, LE CONCLUSIONI DELLA DISCUSSIONE SONO STATE: q AVERE UN CENTRO TIER-1 DEDICATO ALL’ ESPERIMENTO ED UNA DISTRIBUZIONE DI RISORSE FRA DIVERSI TIER-2 E’ APPARSA LA SOLUZIONE MEGLIO RISPONDENTE ALLE ESIGENZE DI ALICE q LA DISTRIBUZIONE OTTIMALE DELLE RISORSE FRA I VARI TIER VERRA’ EFFETTUATA SULLA BASE DELLA SPERIMENTAZIONE q UNA NUOVA DISCUSSIONE SUI TRE MODELLI E’ AL MOMENTO IN ATTO 4/12/00 M. Mazzucato – GR 1 - Roma 18

LOCALIZZAZIONE CENTRI q IPOTESI 1/2 (TIER-1 NON DEDICATO ALL' ESPERIMENTO) q TIER-1 ITALIA q LOCALIZZAZIONE CENTRI q IPOTESI 1/2 (TIER-1 NON DEDICATO ALL' ESPERIMENTO) q TIER-1 ITALIA q IPOTESI 1: CNAF q IPOTESI 2 LNF, LNL q QUESTE DUE IPOTESI SONO AL MOMENTO ALLO STADIO DI DISCUSSIONE PRELIMINARE ALL’ INTERNO DELLA COLLABORAZIONE q CERTAMENTE (cfr. SLIDES PRECEDENTI) IN QUESTI DUE CASI DIVENTA RILEVANTE L’ ORGANIZZAZIONE DEL SUPPORTO SPECIFICO ALL’ ESPERIMENTO, PARTE DEL QUALE, POTREBBE ESSERE SPOSTATO SUI TIER-2 q IPOTESI 3 (TIER-1 DI ESPERIMENTO) q q TIER-1: Torino TIER-2 Bari, Bologna, Catania TIER-3 Cagliari, Catania, Padova, Salerno, Trieste N. B. LE SCELTE DEFINITIVE SUI SITI E SUL BILANCIAMENTO RELATIVO DELLE RISORSE SI PRENDERANNO SULLA BASE DEI RISULTATI DELLA SPERIMENTAZIONE 4/12/00 M. Mazzucato – GR 1 - Roma 21

ATLAS and the INFN RC’s ATLAS views about INFN Tier 1’s q In no ATLAS and the INFN RC’s ATLAS views about INFN Tier 1’s q In no ATLAS INFN site enough manpower resources are available for supporting a Tier 1, as far as system management and operation are concerned. q According to a 1998 estimate, the total manpower of this kind, that could be extracted at LHC start from all the Italian ATLAS sites, amounts to few FTE’s. In no site as much as 2 full FTE’s will be available. q A new recognition will be performed in the next months, taking into account possible reassignment of resources, e. g. from LEP experiments. However, no huge variation is expected. q ATLAS thus need outsourcing of this kind of manpower. We would also accept a solution where INFN provides for the housing, system management and operation of the ATLAS computing h/w in one or more INFN center(s). 4/12/00 M. Mazzucato – GR 1 - Roma 24

ATLAS and the INFN RC’s The manpower for s/w maintenance and user support q ATLAS and the INFN RC’s The manpower for s/w maintenance and user support q ATLAS thinks that the personnel who takes care of the experiment s/w (installation, maintenance, user support etc. ) and some core tools (DB, GRID etc. ) has to work in close contact with the physicists taking care of analysis, simulation, reconstruction, etc. q At least the leading part of this personnel needs to have gained real experience of the experiment s/w, having also normally taken part in the development of some parts. q The personnel with this profile can only be located and trained in the sites where there are groups of physicists actively working in ATLAS s/w and computing. q These requirements apply mostly for the initial phase, after the first years of running there may be less need for the close contact with physicists. 4/12/00 M. Mazzucato – GR 1 - Roma 25

ATLAS and the INFN RC’s The ATLAS Tier 2’s and Tier 3’s q Just ATLAS and the INFN RC’s The ATLAS Tier 2’s and Tier 3’s q Just “desktops” will probably count for 10 -15% of the CPU resources of an experiment. They will be most probably clustered around servers providing common s/w to all the users of a site (Tier 3). In this context, a Tier 2 differs from a Tier 3 mainly because it serves users of different sites. Clearly the mode of operations will evolve with time, starting from one or few prototypes, used by all the groups. q The ATLAS sites who have already expressed interest in participating to the GRID experimentation with the aim of assuming a Tier. N (N>1) role are: q Rome 1, Milan, Naples, Pavia, Rome 2 starting in 2001 q Genova, Pisa starting in 2002 q The s/w experts could cluster around ATLAS Tier 2 sites, which would provide h/w and system resources for testing and implementing the new developments (s/w, GRID, DB etc. ) 4/12/00 M. Mazzucato – GR 1 - Roma 28

ATLAS and the INFN RC’s ATLAS Tier 2’s functionality q As for the functions ATLAS and the INFN RC’s ATLAS Tier 2’s functionality q As for the functions and the h/w resources: q in the next months, we will finalize the layout for the next 3 years (the GRID project span); q for the LHC time, it will be decided in ~1 year from now (in time for the computing Mo. U); 4/12/00 M. Mazzucato – GR 1 - Roma 30

CMS current understandings (1) q CMS Italia preferisce un Tier 1 dove esistano diretti CMS current understandings (1) q CMS Italia preferisce un Tier 1 dove esistano diretti interessi per l’Esperimento q Se cosi’ non dovessere la gestione del Centro diventa “critica” ed inoltre del personale direttamente coinvolto nel Sotware e nelle attivita’ di Calcolo di CMS dovrebbe essere presente al Centro permanentemente q CMS Italia ritiene che il Tier 1 debba essere confrontato con: q Gli altri Tier 1 della Collaborazione q Le dimensioni dei Tier 2 (dislocati presso le Sezioni) q Le funzionalita’ che modificano il Modello originale di MONARC adattandolo/integrandolo con i Tools di GRID q CMS ritene che i Tools alla GRID possano rappresentare un notevole arricchimento delle potenzialita’ del Computing, oltre che del personale coinvolto q Ed infatti CMS e’ fortemente impegnata in questo campo 4/12/00 M. Mazzucato – GR 1 - Roma 38

CMS current understandings (2) q La nuova flessibilta’ introdotta nel Modello Gerarchico di MONARC CMS current understandings (2) q La nuova flessibilta’ introdotta nel Modello Gerarchico di MONARC impone di tenere conto anche dei Tier “inferiori” ai Tier 1. Pertanto non si possono disegnare e programmare soltanto i Tier 1 (ed il Tier 0 al CERN). La regola 1/3 2/3 va correttamente applicata tenedo conto anche delle Istituzioni che non avranno un Tier 1, ma che dovranno poter fare l’Analisi dei dati. q Eseguire l’analisi dei dati in modo efficiente e competitivo e’ il primo ed ultimo scopo del Computing di CMS (ed ovviamente anche della parte italiana di CMS) q Il modello sopra citato (MONARC + GRID) risulta piu’ flessibile ed efficiente nell’utilizzo delle risorse di Computing e di Personale, ed assumono un ruolo “essenziale” i Tier 2, che CMS sta valorizzando a tutti i livelli di funzionalita’ q I Tier 2 permettono una piu’ diretta interazione nell’analisi e trattamento dei dati da parte dei Ricercatori q La “Gerarchia distribuita” che cosi’ viene a crearsi valorizza anche le risorse dei Tier 3 ed persino dei Tier 4 (desktop), permettendo la fattiva collaborazione e completa partecipazione di tutti i Fisici coinvolti, compresi i laurendi! 4/12/00 M. Mazzucato – GR 1 - Roma 39

CMS current understandings (3) q CMS Computing non intende andare in “conflitto” con le CMS current understandings (3) q CMS Computing non intende andare in “conflitto” con le responsabilita’ e gli impegni di personale sui Detector, ma ritiene che l’attivita’ di Computing e la sua realizzazione siano parte integrante di CMS Italia q CMS intende pertanto valutare realistiche implementazioni (ed e’ disponibile a collaborare nella valutazione) di Modelli di Tier 1 e Tier 2 in tempi brevi (qualche mese), valutazioni basate su q Personale disponibile nelle varie sedi e/o da acquisire (anche temporaneo) q Matching delle risorse e delle disponibilita’ di servizi per le attivita’ gia’ oggi in atto (simulazioni e studi di Fisica) q Tempistiche q Infrastrutture necessarie e disponibil q Balancing di funzioni e risorse tra il Tier 1 ed i Tier 2 q Fattibilita’ tecniche e di investimento q Soluzioni di gestione dei centri (compreso l’outsourcing se necessario) q Gestione dei conflitti e della dinamicita’ delle risorse in Tier 1 comuni a piu’ Esperimenti 4/12/00 M. Mazzucato – GR 1 - Roma 40

CMS current understandings (4) q Nell’arco di un anno dovranno essere definiti i compiti, CMS current understandings (4) q Nell’arco di un anno dovranno essere definiti i compiti, le funzionalita’, le dimensioni e le dislocazioni dei Centri (Tier 1 e Tier 2) prototipali per CMS Italia (anche, ma non solo, per arrivare ad un Mo. U sul Computing) q Oggi CMS propone alcune soluzioni temporanee che coprono le esigenze delle “simulazioni” per il prossimo anno (Legnaro, Bari, Bologna, Padova, Pisa, Roma 1, Torino) q CMS intende utilizzare queste soluzioni come “semi” per valutare il progresso e la dislocazione dei prototipi q La scelta delle Sedi, dinamica e progressivamente da discutere coi Referee nominati dall’INFN (oltre che all’interno della Collaborazione), e’ stata gia’ effettuata da CMS, configurando un programma di sviluppo che attraversa tutte le sedi italiane in modo differenziato q CMS ritiene che la scelta di un Tier 1 unico e/o in “condominio” con altri Esperimenti, debba essere effettuata dall’INFN sulla base di un confronto scientifico di progetti differenti, in tempi brevi. 4/12/00 M. Mazzucato – GR 1 - Roma 41

LHCb Computing Model q LHCb data storage q RAW and ESD data stored only LHCb Computing Model q LHCb data storage q RAW and ESD data stored only in production centres (real data at CERN, MC data at the Tier-1 centre which produced them). q No systematic RAW/ESD distribution/replication (very small samples are sent on demand: 10% the first 2 year, 2% later). q No group analyses identified q First stage in the analysis performed in common for all the analyses that subsequently follow. q AOD and TAG produced in production centres (CERN for real data, Tier-1 centrfes for MC). q AOD and TAG data systematically distributed to all Tier-1 centres via network as soon as they are produced (or regenerated) and there stored on disks. q No Tier-2 centres. q 80 TB/a real data AOD exported from CERN. q 120 TB/a MC AOD exported from every Tier-1. 4/12/00 M. Mazzucato – GR 1 - Roma 42

LHCb Regional Centres q LHCb-Italy plans for: q 1 “concentrated” Tier-1 computing centre; q LHCb Regional Centres q LHCb-Italy plans for: q 1 “concentrated” Tier-1 computing centre; q 9 Tier-3 computing centres, located in Bologna, Cagliari, Ferrara, Firenze, Frascati, Genova, Milano, Roma 1, Roma 2. q LHCb Tier-1 can be housed indifferently: q in a LHCb dedicated computing centre (like 2001 LHCb setup in Bologna) provided that technical manpower will be hired and trained by INFN; q in a multi-experiment INFN national computing centre, provided that LHCb will have its own reosurces satisfying the experiment requirements. q LHCb Tier-1 is thought to operate efficiently with resources concentraded in only one site. q LHCb Tier-1 needs a tape library. 4/12/00 M. Mazzucato – GR 1 - Roma 44

VIRGO needs for data analysis Time constraints -The Central Interferometer of VIRGO will produce VIRGO needs for data analysis Time constraints -The Central Interferometer of VIRGO will produce data in 2001 - The full VIRGO interferometer will produce data in 2003 The VIRGO computing model 2 sites for the raw data storage: Tier 0 in Italy and Tier 1 in France Computing for VIRGO in Italy : 1 Tier 0, 2 Tiers 2, 2 Tier 3, Tier functions Tier 0 (raw data storage) Cascina (Virgo site) Tier 2/Tier 3 (data base and computing for pulsar search) Roma/Firenze Tier 2/Tier 3 (coalescent binary system search) Napoli/Perugia 4/12/00 M. Mazzucato – GR 1 - Roma 46

 VIRGO needs for data analysis • Summary of the VIRGO needs for computing VIRGO needs for data analysis • Summary of the VIRGO needs for computing and network connections • • • 4/12/00 units end 2001 CPU capacity+ 8, 000 SI 95 (350 Gflops) estd. number of cpus 400 disk capacity TBytes 10 disk I/O rate GBytes/sec 5 sustained data rate Mbytes/sec 250 WAN links to Cascina Mbits/sec 155 WAN links to labs Mbits/sec 34 WAN links to France Mbits/sec 34 M. Mazzucato – GR 1 - Roma end 2003 8 104 SI 95(3. 5 Tflops) 1000 100 5 250 2 , 500 622 47

VIRGO needs for data analysis Man power for the VIRGO test bed and the VIRGO needs for data analysis Man power for the VIRGO test bed and the full deployment of the system TIER 0: In Cascina there already two system engineers. An operator staff of few units(2/3) is needed to run the system. The European Gravitational Observatory (EGO) consortium can provide it. TIER 2: INFN Section support is under discussion: in case of merging computing necessities from various experiments on site, the man power will be reduced. A realistic alternative for TIER 2 management is the outsourcing: consortia (as for example CASPUR for Roma La Sapienza) can provide the service for TIER 2. 4/12/00 M. Mazzucato – GR 1 - Roma 48

VIRGO needs for data analysis Network improvement is compulsory in any scenario. The IN VIRGO needs for data analysis Network improvement is compulsory in any scenario. The IN 2 P 3 computer center in Lyon already required it explicitly. The Cascina site must be connected to the GARR-B backbone at higher speed immediately. Once the Central Interferometer runs (2001), VIRGO have to start the distribution of the data via network to the VIRGO groups. The analysis of data produced by the network of the Interferometric Gravitational Wave antennas require also network improvements. 4/12/00 M. Mazzucato – GR 1 - Roma 49

CONCLUSIONS q Work to evaluate Tier 1 RC’s location and relation with other Tier(n) CONCLUSIONS q Work to evaluate Tier 1 RC’s location and relation with other Tier(n) in INFN has started q Need time to understand, discuss and agree on possible choices q Cannot be an abstract solution but should be accompanied by a detailed implementation plan defining role and responsibilities of each site q Proposed date to finish the work is around June 2001 4/12/00 M. Mazzucato – GR 1 - Roma 50