ϟ

Antonio Delgado Peris

Here are all the papers by Antonio Delgado Peris that you can download and read on OA.mg.
Antonio Delgado Peris’s last known institution is . Download Antonio Delgado Peris PDFs here.

Claim this Profile →
DOI: 10.1007/s41781-018-0018-8
2019
Cited 121 times
A Roadmap for HEP Software and Computing R&D for the 2020s
Particle physics has an ambitious and broad experimental programme for the coming decades. This programme requires large investments in detector hardware, either to build new facilities and experiments, or to upgrade existing ones. Similarly, it requires commensurate investment in the R&D of software to acquire, manage, process, and analyse the shear amounts of data to be recorded. In planning for the HL-LHC in particular, it is critical that all of the collaborating stakeholders agree on the software goals and priorities, and that the efforts complement each other. In this spirit, this white paper describes the R&D activities required to prepare for this software upgrade.
DOI: 10.1016/j.ijft.2022.100139
2022
Cited 14 times
Numerical analysis of a counter-flow wet cooling tower and its plume
A one-dimensional model to study the heat and mass transfer inside and immediately above a wet counter-flow cooling tower is described. The wet cooling tower thermodynamic model assigns zone-specific Merkel numbers to each of the rain, fill, and spray zones, and it includes an atmospheric plume model. Using the present formulation, zone-by-zone rates of heat rejection and water evaporation can be estimated, as can the visible plume height. The model is validated against the well-established Poppe and Merkel methods as well as select field data. Cooling tower performance and plume visibility are evaluated under a variety of climatic conditions (hot-dry, hot-humid, cool-dry and cool-humid), cooling tower designs (e.g. fill zone height, Hfz), and operating conditions (e.g. water-to-air mass flow rate ratio, L/G). The parametric study in question highlights the ability of the proposed model to predict the impact of ambient conditions, cooling tower design parameters, and operating conditions on overall performance and patterns of atmospheric dispersion. The proposed model is ideal for numerical optimization of cooling towers that need to meet stringent thermal performance and plume visibility requirements.
DOI: 10.1109/tns.2011.2146276
2011
Cited 22 times
CMS Workflow Execution Using Intelligent Job Scheduling and Data Access Strategies
Complex scientific workflows can process large amounts of data using thousands of tasks. The turnaround times of these workflows are often affected by various latencies such as the resource discovery, scheduling and data access latencies for the individual workflow processes or actors. Minimizing these latencies will improve the overall execution time of a workflow and thus lead to a more efficient and robust processing environment. In this paper, we propose a pilot job based infrastructure that has intelligent data reuse and job execution strategies to minimize the scheduling, queuing, execution and data access latencies. The results have shown that significant improvements in the overall turnaround time of a workflow can be achieved with this approach. The proposed approach has been evaluated, first using the CMS Tier0 data processing workflow, and then simulating the workflows to evaluate its effectiveness in a controlled environment.
DOI: 10.1088/1742-6596/2438/1/012039
2023
Extending the distributed computing infrastructure of the CMS experiment with HPC resources
Abstract Particle accelerators are an important tool to study the fundamental properties of elementary particles. Currently the highest energy accelerator is the LHC at CERN, in Geneva, Switzerland. Each of its four major detectors, such as the CMS detector, produces dozens of Petabytes of data per year to be analyzed by a large international collaboration. The processing is carried out on the Worldwide LHC Computing Grid, that spans over more than 170 compute centers around the world and is used by a number of particle physics experiments. Recently the LHC experiments were encouraged to make increasing use of HPC resources. While Grid resources are homogeneous with respect to the used Grid middleware, HPC installations can be very different in their setup. In order to integrate HPC resources into the highly automatized processing setups of the CMS experiment a number of challenges need to be addressed. For processing, access to primary data and metadata as well as access to the software is required. At Grid sites all this is achieved via a number of services that are provided by each center. However at HPC sites many of these capabilities cannot be easily provided and have to be enabled in the user space or enabled by other means. At HPC centers there are often restrictions regarding network access to remote services, which is again a severe limitation. The paper discusses a number of solutions and recent experiences by the CMS experiment to include HPC resources in processing campaigns.
DOI: 10.48550/arxiv.2404.02100
2024
Analysis Facilities White Paper
This white paper presents the current status of the R&D for Analysis Facilities (AFs) and attempts to summarize the views on the future direction of these facilities. These views have been collected through the High Energy Physics (HEP) Software Foundation's (HSF) Analysis Facilities forum, established in March 2022, the Analysis Ecosystems II workshop, that took place in May 2022, and the WLCG/HSF pre-CHEP workshop, that took place in May 2023. The paper attempts to cover all the aspects of an analysis facility.
DOI: 10.1051/epjconf/202429507027
2024
Integration of the Barcelona Supercomputing Center for CMS computing: Towards large scale production
The CMS experiment is working to integrate an increasing number of High Performance Computing (HPC) resources into its distributed computing infrastructure. The case of the Barcelona Supercomputing Center (BSC) is particularly challenging as severe network restrictions prevent the use of CMS standard computing solutions. The CIEMAT CMS group has performed significant work in order to overcome these constraints and make BSC resources available to CMS. The developments include adapting the workload management tools, replicating the CMS software repository to BSC storage, providing an alternative access to detector conditions data, and setting up a service to transfer produced output data to a nearby storage facility. In this work, we discuss the current status of this integration activity and present recent developments, such as a front-end service to improve slot usage efficiency and an enhanced transfer service that supports the staging of input data for workflows at BSC. Moreover, significant efforts have been devoted to improving the scalability of the deployed solution, automating its operation, and simplifying the matchmaking of CMS workflows that are suitable for execution at BSC.
DOI: 10.1051/epjconf/202429501006
2024
A case study of content delivery networks for the CMS ex-periment
In 2029 the LHC will start the high-luminosity LHC program, with a boost in the integrated luminosity resulting in an unprecedented amount of ex- perimental and simulated data samples to be transferred, processed and stored in disk and tape systems across the worldwide LHC computing Grid. Content de- livery network solutions are being explored with the purposes of improving the performance of the compute tasks reading input data via the wide area network, and also to provide a mechanism for cost-effective deployment of lightweight storage systems supporting traditional or opportunistic compute resources. In this contribution we study the benefits of applying cache solutions for the CMS experiment, in particular the configuration and deployment of XCache serving data to two Spanish WLCG sites supporting CMS: the Tier-1 site at PIC and the Tier-2 site at CIEMAT. The deployment and configuration of the system and the developed monitoring tools will be shown, as well as data popularity studies in relation to the optimization of the cache configuration, the effects on CPU efficiency improvements for analysis tasks, and the cost benefits and impact of including this solution in the region.
DOI: 10.1051/epjconf/202429507045
2024
The Spanish CMS Analysis Facility at CIEMAT
The increasingly larger data volumes that the LHC experiments will accumulate in the coming years, especially in the High-Luminosity LHC era, call for a paradigm shift in the way experimental datasets are accessed and analyzed. The current model, based on data reduction on the Grid infrastructure, followed by interactive data analysis of manageable size samples on the physicists’ individual computers, will be superseded by the adoption of Analysis Facilities. This rapidly evolving concept is converging to include dedicated hardware infrastructures and computing services optimized for the effective analysis of large HEP data samples. This paper describes the actual implementation of this new analysis facility model at the CIEMAT institute, in Spain, to support the local CMS experiment community. Our work details the deployment of dedicated highly performant hardware, the operation of data staging and caching services ensuring prompt and efficient access to CMS physics analysis datasets, and the integration and optimization of a custom analysis framework based on ROOT’s RDataFrame and CMS NanoAOD format. Finally, performance results obtained by benchmarking the deployed infrastructure and software against a CMS analysis workflow are summarized.
DOI: 10.1007/s12083-015-0338-y
2015
Cited 8 times
Evaluation of alternatives for the broadcast operation in Kademlia under churn
DOI: 10.1109/hpcsim.2014.6903678
2014
Cited 6 times
Distributed scheduling and data sharing in late-binding overlays
Pull-based late-binding overlays are used in some of today's largest computational grids. Job agents are submitted to resources with the duty of retrieving real workload from a central queue at runtime. This helps overcome the problems of these complex environments: heterogeneity, imprecise status information and relatively high failure rates. In addition, the late job assignment allows dynamic adaptation to changes in grid conditions or user priorities. However, as the scale grows, the central assignment queue may become a bottleneck for the whole system. This article presents a distributed scheduling architecture for late-binding overlays, which addresses this issue by letting execution nodes build a distributed hash table and delegating job matching and assignment to them. This reduces the load on the central server and makes the system much more scalable and robust. Scalability makes fine-grained scheduling possible and enables new functionalities, like the implementation of a distributed data cache on the execution nodes, which helps alleviate the commonly congested grid storage services.
DOI: 10.1109/hpcc.2012.107
2012
Cited 6 times
Evaluation of the Broadcast Operation in Kademlia
Several proposals exist that try to enhance Distributed Hash Table (DHT) systems with broadcasting capabilities. None of them however specifically addresses the particularities of Kademlia, an important DHT, used in well known real applications. Our work analyzes the implications of Kademlia's use of XOR-based distance metrics and subsequently discusses the applicability of existing broadcasting proposals to it. Based on this, several algorithms for broadcasting in Kademlia have been implemented and experimentally evaluated under different conditions of churn and failure rate. All significant assessment criteria have been considered: node coverage, messages to nodes ratio, latency and imbalance factor. Since no perfect solution exists, a discussion on the choices and compromises to make depending on system characteristics or application priorities is presented. In addition, several enhancements are proposed that profit from Kademlia characteristics in order to make the broadcasting more robust against stale routing information or malfunctioning nodes.
2019
Cited 6 times
A roadmap for HEP software and computing R&D for the 2020s
Particle physics has an ambitious and broad experimental programme for the coming decades. This programme requires large investments in detector hardware, either to build new facilities and experiments, or to upgrade existing ones. Similarly, it requires commensurate investment in the R&D of software to acquire, manage, process, and analyse the shear amounts of data to be recorded. In planning for the HL-LHC in particular, it is critical that all of the collaborating stakeholders agree on the software goals and priorities, and that the efforts complement each other. In this spirit, this white paper describes the R&D activities required to prepare for this software upgrade.
DOI: 10.1088/1742-6596/219/6/062043
2010
Cited 6 times
Data location-aware job scheduling in the grid. Application to the GridWay metascheduler
Grid infrastructures constitute nowadays the core of the computing facilities of the biggest LHC experiments. These experiments produce and manage petabytes of data per year and run thousands of computing jobs every day to process that data. It is the duty of metaschedulers to allocate the tasks to the most appropriate resources at the proper time.
DOI: 10.1007/s10723-016-9374-y
2016
Distributed Late-binding Scheduling and Cooperative Data Caching
DOI: 10.1016/j.nuclphysbps.2007.11.128
2008
Cited 3 times
Data Location, Transfer and Bookkeeping in CMS
Within 5 years CMS expects to be managing many tens of petabytes of data at over a hundred sites around the world. This represents more than an order of magnitude increase in data volume over existing HEP experiments. The underlying concepts and architecture of the CMS model for distributed data management will be presented. The technical descriptions of the main data management components for data transfer, dataset bookkeeping, data location and file access will be described. In addition we will present the experience in using the system in CMS data challenges and ongoing MC production.
DOI: 10.5170/cern-2005-002.715
2005
Cited 3 times
Experience integrating a General Information System API in LCG Job Management and Monitoring Services
DOI: 10.1051/epjconf/202024509007
2020
Exploiting network restricted compute resources with HTCondor: a CMS experiment experience
In view of the increasing computing needs for the HL-LHC era, the LHC experiments are exploring new ways to access, integrate and use non-Grid compute resources. Accessing and making efficient use of Cloud and High Performance Computing (HPC) resources present a diversity of challenges for the CMS experiment. In particular, network limitations at the compute nodes in HPC centers prevent CMS pilot jobs to connect to its central HTCondor pool in order to receive payload jobs to be executed. To cope with this limitation, new features have been developed in both HTCondor and the CMS resource acquisition and workload management infrastructure. In this novel approach, a bridge node is set up outside the HPC center and the communications between HTCondor daemons are relayed through a shared file system. This conforms the basis of the CMS strategy to enable the exploitation of the Barcelona Supercomputing Center (BSC) resources, the main Spanish HPC site. CMS payloads are claimed by HTCondor condor_startd daemons running at the nearby PIC Tier-1 center and routed to BSC compute nodes through the bridge. This fully enables the connectivity of CMS HTCondor-based central infrastructure to BSC resources via the PIC HTCondor pool. Other challenges include building custom singularity images with CMS software releases, bringing conditions data to payload jobs, and custom data handling between BSC and PIC. This report describes the initial technical prototype, its deployment and tests, and future steps. A key aspect of the technique described in this contribution is that it could be universally employed in similar network-restrictive HPC environments elsewhere.
DOI: 10.1051/epjconf/202024503013
2020
Lightweight site federation for CMS support
There is a general trend in WLCG towards the federation of resources, aiming for increased simplicity, efficiency, flexibility, and availability. Although general VO-agnostic federation of resources between two independent and autonomous resource centres may prove arduous, a considerable amount of flexibility in resource sharing can be achieved in the context of a single WLCG VO, with a relatively simple approach. We have demonstrated this for PIC and CIEMAT, the Spanish Tier-1 and Tier-2 sites for CMS, by making use of the existing CMS xrootd federation infrastructure and profiting from the common CE/batch technology used by the two centres. This work describes how compute slots are shared between the two sites, so that the capacity of one site can be dynamically increased with idle execution slots from the remote site, and how data can be efficiently accessed irrespective of its location. Our contribution includes measurements for diverse CMS workflows comparing performances between local and remote execution, and can also be regarded as a benchmark to explore future potential scenarios, where storage resources would be concentrated in a reduced number of sites.
DOI: 10.5170/cern-2005-002.986
2005
Toward a Grid Technology Independent Programming Interface for HEP Applications
DOI: 10.1051/epjconf/202024504028
2020
CMS data access and usage studies at PIC Tier-1 and CIEMAT Tier-2
The current computing models from LHC experiments indicate that much larger resource increases would be required by the HL-LHC era (2026+) than those that technology evolution at a constant budget could bring. Since worldwide budget for computing is not expected to increase, many research activities have emerged to improve the performance of the LHC processing software applications, as well as to propose more efficient resource deployment scenarios and data management techniques, which might reduce this expected increase of resources. The massively increasing amounts of data to be processed leads to enormous challenges for HEP storage systems, networks and the data distribution to end-users. These challenges are particularly important in scenarios in which the LHC data would be distributed from small numbers of centers holding the experiment’s data. Enabling data locality relative to computing tasks via local caches on sites seems a very promising approach to hide transfer latencies while reducing the deployed storage space and number of replicas overall. However, this highly depends on the workflow I/O characteristics and available network across sites. A crucial assessment of how the experiments are accessing and using the storage services deployed in WLCG sites, to evaluate and simulate the benefits for several of the new emerging proposals within WLCG/HSF. Studies on access and usage of storage, data access and popularity studies for the CMS workflows executed in the Spanish Tier-1 (PIC) and Tier-2 (CIEMAT) sites supporting CMS activities are reviewed in this report, based on local and experiment monitoring data, spanning more than one year. This is of relevance for simulation of data caches for end-user analysis data, as well as identifying potential areas for storage savings.
DOI: 10.5170/cern-2005-002.719
2005
Experiment Software Installation experience in LCG-2
2015
Microplanificación de asignación tardía y almacenamiento temporal distribuidos para flujos de trabajo intensivos en datos
El mundo de hoy en dia se encuentra inundado por ingentes cantidades de informacion digital procedente de muy diversas fuentes. Todo apunta, ademas, a que esta tendencia se agudizara en el futuro. Ni la industria, ni la sociedad en general, ni, muy particularmente, la ciencia, permanecen indiferentes ante este hecho. Al contrario, se esfuerzan por obtener el maximo provecho de esta informacion, lo que significa que deben capturarla, transferirla, almacenarla y procesarla puntual y eficientemente, utilizando una amplia gama de recursos computacionales. Pero esta tarea no es siempre sencilla. Un ejemplo representativo de los desafios que suponen el manejo y procesamiento de grandes cantidades de datos es el de los experimentos de fisica de particulas del Large Hadron Collider (LHC), en Ginebra, que cada ano deben gestionar decenas de petabytes de informacion. Basandonos en la experiencia de una de estas colaboraciones, hemos estudiado los principales problemas relativos a la gestion de volumenes de datos masivos y a la ejecucion de vastos flujos de trabajo que necesitan consumirlos. En este contexto, hemos desarrollado una arquitectura de proposito general para la planificacion y ejecucion de flujos de trabajo con importantes requisitos de datos, que hemos llamado Task Queue. Este nuevo sistema aprovecha el modelo de asignacion tardia basado en agentes que ha ayudado a los experimentos del LHC a superar los problemas asociados con la heterogeneidad y la complejidad de las grandes infraestructuras grid de computacion. Nuestra propuesta presenta varias mejoras con respecto a los sistemas existentes. Los agentes de ejecucion de la arquitectura Task Queue comparten una tabla hash distribuida (Distributed Hash Table, DHT) y realizan la asignacion de tareas de una manera cooperativa. De esta forma, se evitan los problemas de escalabilidad de los algoritmos centralizados de asignacion y se mejoran los tiempos de ejecucion. Esta escalabilidad nos permite realizar una microplanificacion de grano fino lo cual posibilita nuevas funcionalidades, como la implementacion de una cache distribuida en los nodos de ejecucion y el uso de la informacion de ubicacion de los datos en las decisiones de asignacion de tareas. Esto mejora la eficiencia del procesado de datos y ayuda a aliviar los habitualmente congestionados servicios de almacenamiento del grid. Ademas, nuestro sistema es mas robusto frente a problemas en la interaccion con la cola central de tareas y ofrece mejor comportamiento en situaciones con patrones de acceso a datos exigentes o en ausencia de servicios de almacenamiento locales. Todo esto ha sido demostrado en una amplia serie de pruebas de evaluacion. Dado que nuestro procedimiento de planificacion de tareas distribuido requiere el uso de mensajes de broadcast, tambien hemos realizado un profundo estudio de las posibles aproximaciones a la implementacion de esta operacion sobre el DHT Kademlia, el cual es utilizado para la cache de datos compartida. Kademlia ofrece enrutamiento a nodos individuales pero no incluye ninguna primitiva de broadcast. Nuestro trabajo expone las peculiaridades de este sistema, particularmente su metrica basada en la operacion XOR, y estudia analiticamente que tecnicas de broadcast pueden ser usadas con el. Tambien se ha desarrollado un modelo que estima la cobertura de nodos en funcion de la probabilidad que cada mensaje individual alcance su destino correctamente. Como validacion, los algoritmos se han implementado y se han evaluado exhaustivamente. Ademas, proponemos varias tecnicas para mejorar los protocolos en situaciones adversas, por ejemplo cuando el sistema presenta una alta rotacion de nodos o la tasa de error en las entregas no es despreciable. Esta tecnicas incluyen redundancia, reenvio e inundacion (flooding), asi como combinaciones de las mismas. Presentamos un analisis de las fortalezas y debilidades de los diferentes algoritmos y las mencionadas tecnicas complementarias.
DOI: 10.1088/1742-6596/119/5/052008
2008
Exercising CMS dataflows and workflows in computing challenges at the SpanishTier-1 and Tier-2 sites
An overview of the data transfer, processing and analysis operations conducted at the Spanish Tier-1 (PIC, Barcelona) and Tier-2 (CIEMAT-Madrid and IFCA-Santander federation) centres during the past CMS CSA06 Computing, Software and Analysis challenge and in preparation for CSA07 is presented.
DOI: 10.48550/arxiv.2212.04889
2022
Second Analysis Ecosystem Workshop Report
The second workshop on the HEP Analysis Ecosystem took place 23-25 May 2022 at IJCLab in Orsay, to look at progress and continuing challenges in scaling up HEP analysis to meet the needs of HL-LHC and DUNE, as well as the very pressing needs of LHC Run 3 analysis. The workshop was themed around six particular topics, which were felt to capture key questions, opportunities and challenges. Each topic arranged a plenary session introduction, often with speakers summarising the state-of-the art and the next steps for analysis. This was then followed by parallel sessions, which were much more discussion focused, and where attendees could grapple with the challenges and propose solutions that could be tried. Where there was significant overlap between topics, a joint discussion between them was arranged. In the weeks following the workshop the session conveners wrote this document, which is a summary of the main discussions, the key points raised and the conclusions and outcomes. The document was circulated amongst the participants for comments before being finalised here.
DOI: 10.5281/zenodo.7418264
2022
HSF IRIS-HEP Second Analysis Ecosystem Workshop Report
DOI: 10.1051/epjconf/202125102020
2021
Exploitation of network-segregated CPU resources in CMS
CMS is tackling the exploitation of CPU resources at HPC centers where compute nodes do not have network connectivity to the Internet. Pilot agents and payload jobs need to interact with external services from the compute nodes: access to the application software (CernVM-FS) and conditions data (Frontier), management of input and output data files (data management services), and job management (HTCondor). Finding an alternative route to these services is challenging. Seamless integration in the CMS production system without causing any operational overhead is a key goal. The case of the Barcelona Supercomputing Center (BSC), in Spain, is particularly challenging, due to its especially restrictive network setup. We describe in this paper the solutions developed within CMS to overcome these restrictions, and integrate this resource in production. Singularity containers with application software releases are built and pre-placed in the HPC facility shared file system, together with conditions data files. HTCondor has been extended to relay communications between running pilot jobs and HTCondor daemons through the HPC shared file system. This operation mode also allows piping input and output data files through the HPC file system. Results, issues encountered during the integration process, and remaining concerns are discussed.
1997
Effects of different UV exposures on the thermoluminescence of high albite
High albite can be used as a new UV dosimeter. Heated natural albite posseses a special property: it has strong thermoluminescence (TL) blue emissions under UV radiation. The TL glow curves of high albite after different times of UV exposure show PhotoTransferred ThermoLuminescence (PTTL) from deeps traps (345C 450C) to shallow traps (150C) which can be used to store UV damage. The point of inflexion of the process occurs after one hour under UV radiation, the low temperature peaks start gaining in intensity while the high temperature peaks stop dropping and stabilize into a plateau. Samples which have been subjected to UV radiation for 1 hour at different temperatures (RT, 70C, 105C and 140C) give glow curves which display an increase of the maxima peaks according to temperature (up to 300C). High albite, UV irradiated (for one hour) at 70C displays a TL signal decrease of 30% after 1000 hours of storage at room temperature with high intensity emissions, for this reason this material could be a good UV dosimeter.
DOI: 10.5209/rev_cfcl.1998.v15.35801
1998
La lista de Voces animantium del Matritensis B. N. 19: estudio de sus fuentes y nueva edición.
DOI: 10.5209/rev_cfcl.1996.v11.35875
1996
El Ritmo de Roncesvalles: estudio y edición.
1988
The tunneling effect between the blue and UV luminescence spectra bands in preheated adularia.
DOI: 10.5209/rev_cfcl.1992.v3.36019
1992
Notas para una mejora de la traducción española de la Introducción al latín vulgar de Veikko Väänänen
1991
[Toscana region: experimental phase of rescue with helicopter].