The post Three years of building bridges in HPC research between Europe and Latin America: RISC2 project comes to an end first appeared on RISC2 Project.
]]>“The RISC2 project has proven to be a team effort in which European and Latin American partners worked together to drive HPC collaboration forward. We have been able to create a lively and active community across the Atlantic to stimulate dialogue and boost cooperation that won’t die with RISC2’s formal end”, says Fabrizio Gagliardi, manager director of RISC2.
Since 2021, this knowledge-sharing network has organised webinars, summer schools, meetings with policymakers and participated in conferences and dissemination events on both sides of the Atlantic. The project also resulted in the HPC Observatory Repository — a collection of documents and training materials produced as part of the project – and the White Paper on HPC R&I Collaboration Opportunities, a document that reviews the key socio-economic and environmental factors and trends that influence HPC needs.
These were two of the issues highlighted by European Commission officials and experts during the final evaluation of the project, which could provide continuity to the work carried out by the consortium over the last three years, in line with the wishes of the partners and the advice of the evaluators. “Beyond RISC2, we should keep the momentum and leverage the importance of Latin America in the frame of the Green Deal actions: HPC stakeholders should encourage policymakers to build bilateral agreements and offer open calls focused on HPC collaboration“, reflects Fabrizio Gagliardi.
The post Three years of building bridges in HPC research between Europe and Latin America: RISC2 project comes to an end first appeared on RISC2 Project.
]]>The post CARLA 2023: RISC2 Results Presented at Largest HPC Conference in Latin America first appeared on RISC2 Project.
]]>Given the nature of the project, RISC2 could not fail to be represented through its partners and with a strong presence in the event’s program. Carlos J. Barrios, researcher from the Universidad Industrial de Santander and RISC2 partner was responsible for opening CARLA, with his address setting the tone for the conference, emphasizing the importance of collaborative efforts and knowledge sharing in furthering the frontiers of HPC.
Fabrizio Gagliardi, the coordinator of RISC2, also took center stage with a special talk that introduced the audience to the mission and objectives of the RISC2 project. The presentation shed light on the pivotal role that RISC2 plays in advancing HPC research and development of the cooperation between the two continents in this field. Gagliardi participated in the EuroHPCLatam panel: Policy and Global Actions, which included representatives from Red Clara, CAF and the Ministry of CyT Colombia. This panel explored the policies and global actions required to propel HPC forward in Latin America, emphasizing collaboration between key stakeholders.
Another highlight of CARLA 2023 was the tribute to Mateo Valero, one of the promoters of RISC2. Valero’s dedication and contributions to the field were celebrated through an award with his name and one he was the first recepient, underscoring the lasting impact of his work on the entire HPC community.
This event was particularly important as it coincided with the end of the RISC2 project and the presentation of its results. Over the course of three years, the initiative has strengthened contacts and promoted the exchange of knowledge between researchers from Latin America and Europe through the organization of nine webinars, the support of several schools, workshops, and other training events in the field for young students and researchers. During this period, RISC2 partners also participated in several conferences and ceremonies with policymakers to raise awareness of the importance of continuing to support and prioritize this area of research in the future.
The post CARLA 2023: RISC2 Results Presented at Largest HPC Conference in Latin America first appeared on RISC2 Project.
]]>The post RISC2’s partners gather in Brussels to reflect on three years of collaboration between EU and Latin America first appeared on RISC2 Project.
]]>The session began with a welcome and introduction by Mateo Valero (BSC), one of the main drivers of this cooperation and a leading name in the field of HPC. This intervention was later complemented by Fabrizio Gagliardi (BSC). Afterward, Elsa Carvalho (INESC TEC) presented the work done in terms of communication by the RISC2 team, an important segment for all the news and achievements to reach all the partners and countries involved.
Carlos J. Barrios Hernandez then presented the work done within the HPC Observatory, a relevant source of information that European and Latin American research organizations can address with HPC and/or AI issues.
The session closed with an important and pertinent debate on how to strengthen cooperation in HPC between the European Union and Latin America, in which all participants contributed and gave their opinion, committing to efforts so that the work developed within the framework of RISC2 is continued.
What our partners had to say about the meeting?
Rafael Mayo Garcia, CIEMAT:
“The policy event organized by RISC2 in Brussels was of utmost importance for the development of HPC and digital capabilities for a shared infrastructure between EU and LAC. Even more, it has had crucial contributions to international entities such as CYTED, the Ibero-American Programme for the Development of Science and Technology. On the CIEMAT side, it has been a new step beyond for building and participating in a HPC shared ecosystem.”
Esteban Meneses, CeNAT:
“In Costa Rica, CeNAT plays a critical role in fostering technological change. To achieve that goal, it is fundamental to synchronize our efforts with other key players, particularly government institutions. The event policy in Brussels was a great opportunity to get closer to our science and technology ministry and start a dialogue on the importance of HPC, data science, and artificial intelligence for bringing about the societal changes we aim for.”
Esteban Mocskos, UBA:
“The Policy Event recently held in Brussels and organized by the RISC2 project had several remarkable points. The gathering of experts in HPC research and management in Latin America and Europe served to plan the next steps in the joint endeavor to deepen the collaboration in this field. The advance in management policies, application optimization, and user engagement are fundamental topics treated during the main sessions and also during the point-to-point talks in every corner of the meeting room.
I can say that this meeting will also spawn different paths in these collaboration efforts that we’ll surely see their results during the following years with a positive impact on both sides of this fruitful relationship: Latin America and Europe.”
Sergio Nesmachnow, Universidad de la República:
“The National Supercomputing Center (Uruguay) and Universidad de la República have led the development of HPC strategies and technologies and their application to relevant problems in Uruguay. Specific meetings such as the policy event organized by RISC2 in Brussels are key to present and disseminate the current developments and achievements to relevant political and technological leaders in our country, so that they gain knowledge about the usefulness of HPC technologies and infrastructure to foster the development of national scientific research in capital areas such as sustainability, energy, and social development. It was very important to present the network of collaborators in Latin America and Europe and to show the involvement of institutional and government agencies.
Within the contacts and talks during the organization of the meeting, we introduced the projecto to national authorities, including the National Director of Science and Technology, Ministry of Education and Culture, and the President of the National Agency for Research and Innovation, as well as the Uruguayan Agency for International Cooperation and academic authorities from all institutions involved in the National Supercomputing Center initiative. We hope the established contacts can result in productive joint efforts to foster the development of HPC and related scientific areas in our country and the region.”
Carla Osthoff, LNCC:
“In Brazil, LNCC is critical in providing High Performance Computing Resources for the Research Community and training Human Resources and fostering new technologies. The policy event organized by RISC2 in Brussels was fundamental to synchronizing LNCC efforts with other government institutions and international entities. On the LNCC side, it has been a new step beyond building and participating in an HPC-shared ecosystem.
Specific meetings such as the policy event organized by RISC2 in Brussels were very important to present the network of collaborators in Latin America and Europe and to show the involvement of institutional and government agencies.
As a result of joint activities in research and development in the areas of information and communication technologies (ICT), artificial intelligence, applied mathematics, and computational modelling, with emphasis on the areas of scientific computing and data science, a Memorandum of Understanding (MoU) have been signed between LNCC and Inria/France. As a result of new joint activities, LNCC and INESC TEC/Portugal are starting collaboration through INESC TEC International Visiting Researcher Programme 2023.”
The post RISC2’s partners gather in Brussels to reflect on three years of collaboration between EU and Latin America first appeared on RISC2 Project.
]]>The post Future of EU-LATAM collaboration on HPC first appeared on RISC2 Project.
]]>Read more here.
Check the agenda:
The post Future of EU-LATAM collaboration on HPC first appeared on RISC2 Project.
]]>The post Subsequent Progress And Challenges Concerning The México-UE Project ENERXICO: Supercomputing And Energy For México first appeared on RISC2 Project.
]]>The ENERXICO Project focused on developing advanced simulation software solutions for oil & gas, wind energy and transportation powertrain industries. The institutions that collaborated in the project are for México: ININ (Institution responsible for México), Centro de Investigación y de Estudios Avanzados del IPN (Cinvestav), Universidad Nacional Autónoma de México (UNAM IINGEN, FCUNAM), Universidad Autónoma Metropolitana-Azcapotzalco, Instituto Mexicano del Petróleo, Instituto Politécnico Nacional (IPN) and Pemex, and for the European Union: Centro de Supercómputo de Barcelona (Institution responsible for the EU), Technische Universitäts München, Alemania (TUM), Universidad de Grenoble Alpes, Francia (UGA), CIEMAT, España, Repsol, Iberdrola, Bull, Francia e Universidad Politécnica de Valencia, España.
The Project contemplated four working packages (WP):
WP1 Exascale Enabling: This was a cross-cutting work package that focused on assessing performance bottlenecks and improving the efficiency of the HPC codes proposed in vertical WP (UE Coordinator: BULL, MEX Coordinator: CINVESTAV-COMPUTACIÓN);
WP2 Renewable energies: This WP deployed new applications required to design, optimize and forecast the production of wind farms (UE Coordinator: IBR, MEX Coordinator: ININ);
WP3 Oil and gas energies: This WP addressed the impact of HPC on the entire oil industry chain (UE Coordinator: REPSOL, MEX Coordinator: ININ);
WP4 Biofuels for transport: This WP displayed advanced numerical simulations of biofuels under conditions similar to those of an engine (UE Coordinator: UPV-CMT, MEX Coordinator: UNAM);
For WP1 the following codes were optimized for exascale computers: Alya, Bsit, DualSPHysics, ExaHyPE, Seossol, SEM46 and WRF.
As an example, we present some of the results for the DualPHYysics code. We evaluated two architectures: The first set of hardware used were identical nodes, each equipped with 2 ”Intel Xeon Gold 6248 Processors”, clocking at 2.5 GHz with about 192 GB of system memory. Each node contained 4 Nvidia V100 Tesla GPUs with 32 GB of main memory each. The second set of hardware used were identical nodes, each equipped with 2 ”AMD Milan 7763 Processors”, clocking at 2.45 GHz with about 512 GB of system memory. Each node contained 4 Nvidia V100 Ampere GPUs with 40 GB of main memory each. The code was compiled and linked with CUDA 10.2 and OpenMPI 4. The application was executed using one GPU per MPI rank.
In Figures 1 and 2 we show the scalability of the code for the strong and weak scaling tests that indicate that the scaling is very good. Motivated by these excellent results, we are in the process of performing in the LUMI supercomputer new SPH simulations with up to 26,834 million particles that will be run with up to 500 GPUs, which is 53.7 million particles per GPU. These simulations will be done initially for a Wave Energy Converter (WEC) Farm (see Figure 3), and later for turbulent models.
Figure 1. Strong scaling test with a fix number of particles but increasing number of GPUs.
Figure 2. Weak scaling test with increasing number of particles and GPUs.
Figure 3. Wave Energy Converter (WEC) Farm (taken from https://corpowerocean.com/)
As part of WP3, ENERXICO developed a first version of a computer code called Black Hole (or BH code) for the numerical simulation of oil reservoirs, based on the numerical technique known as Smoothed Particle Hydrodynamics or SPH. This new code is an extension of the DualSPHysics code (https://dual.sphysics.org/) and is the first SPH based code that has been developed for the numerical simulation of oil reservoirs and has important benefits versus commercial codes based on other numerical techniques.
The BH code is a large-scale massively parallel reservoir simulator capable of performing simulations with billions of “particles” or fluid elements that represent the system under study. It contains improved multi-physics modules that automatically combine the effects of interrelated physical and chemical phenomena to accurately simulate in-situ recovery processes. This has led to the development of a graphical user interface, considered as a multiple-platform application for code execution and visualization, and for carrying out simulations with data provided by industrial partners and performing comparisons with available commercial packages.
Furthermore, a considerable effort is presently being made to simplify the process of setting up the input for reservoir simulations from exploration data by means of a workflow fully integrated in our industrial partners’ software environment. A crucial part of the numerical simulations is the equation of state. We have developed an equation of state based on crude oil data (the so-called PVT) in two forms, the first as a subroutine that is integrated into the code, and the second as an interpolation subroutine of properties’ tables that are generated from the equation of state subroutine.
An oil reservoir is composed of a porous medium with a multiphase fluid made of oil, gas, rock and other solids. The aim of the code is to simulate fluid flow in a porous medium, as well as the behaviour of the system at different pressures and temperatures. The tool should allow the reduction of uncertainties in the predictions that are carried out. For example, it may answer questions about the benefits of injecting a solvent, which could be CO2, nitrogen, combustion gases, methane, etc. into a reservoir, and the times of eruption of the gases in the production wells. With these estimates, it can take the necessary measures to mitigate their presence, calculate the expense, the pressure to be injected, the injection volumes and most importantly, where and for how long. The same happens with more complex processes such as those where fluids, air or steam are injected, which interact with the rock, oil, water and gas present in the reservoir. The simulator should be capable of monitoring and preparing measurement plans.
In order to be able to perform a simulation of a reservoir oil field, an initial model needs to be created. Using geophysical forward and inverse numerical techniques, the ENERXICO project evaluated novel, high-performance simulation packages for challenging seismic exploration cases that are characterized by extreme geometric complexity. Now, we are undergoing an exploration of high-order methods based upon fully unstructured tetrahedral meshes and also tree-structured Cartesian meshes with adaptive mesh refinement (AMR) for better spatial resolution. Using this methodology, our packages (and some commercial packages) together with seismic and geophysical data of naturally fractured reservoir oil fields, are able to create the geometry (see Figure 4), and exhibit basic properties of the oil reservoir field we want to study. A number of numerical simulations are performed and from these oil fields exploitation scenarios are generated.
Figure 4. A detail of the initial model for a SPH simulation of a porous medium.
More information about the ENERXICO Project can be found in: https://enerxico-project.eu/
By: Jaime Klapp (ININ, México) and Isidoro Gitler (Cinvestav, México)
The post Subsequent Progress And Challenges Concerning The México-UE Project ENERXICO: Supercomputing And Energy For México first appeared on RISC2 Project.
]]>The post More than 100 students participated in the HPC, Data & Architecture Week first appeared on RISC2 Project.
]]>This event had four main courses: “Foundations of Parallel Programming”, “Large scale data processing and machine learning”, “New architectures and specific computing platforms”, and “Administrations techniques for large-scale computing facilities”.
More than 100 students actively participated in the event who traveled from different part of the country. 30 students received financial support to participate (traveling and living) provided by the National HPC System (SNCAD) dependent of the Argentina’s Ministry of Science.
Esteban Mocskos, one of the organizers of the event, believes “this kind of events should be organized regularly to sustain the flux of students in the area of HPC”. In his opinion, “a lot of students from Argentina get their first contact with HPC topics. As such a large country, impacting a distant region also means impacting the neighboring countries. Those students will bring their experience to other students in their places”. According to Mocskos, initiatives like the “HPC, Data & Architecture Week” spark a lot of collaborations.
The post More than 100 students participated in the HPC, Data & Architecture Week first appeared on RISC2 Project.
]]>The post Towards a greater HPC capacity in Latin America first appeared on RISC2 Project.
]]>A country that does not have the computational capacity to solve its own problems will have no alternative but to try to acquire solutions provided by others. One of the most important aspects of sovereignty in the 21st century is the ability to produce mathematical models and to have the capacity to solve them. Today, the availability of computing power commensurate with one’s wealth exponentially increases a country’s capacity to produce knowledge. in the developed world, it is estimated that for every dollar invested in supercomputing, the return to society is of the order of US$ 44(1) and to the academic world US$ 30(2). For these reasons, HPC occupies an important place on the political and diplomatic agendas of developed countries.
In Latin America, investment in HPC is very low compared to what’s the US, Asia and Europe are doing. In order to quantify this difference, we present the tables below, which show the accumulated computing capacity in the ranking of the 500 most powerful supercomputers in the world – the TOP500(3) – (Table 1), and the local reality (Table 2). Other data are also included, such as the population (in millions), the number of researchers per 1,000 inhabitants (Res/1000), the computing capacity per researcher (Gflops/Res) and the computing capacity per US$ million of GPD. In Table 1, we have grouped the countries by geographical area. America appears as the area with the highest computing capacity, essentially due to the USA, which has almost 45% of the world’s computing capacity in the TOP500. It if followed by Asia and then Europe. Tis TOP500 list includes mainly academic research centres, but also industry ones, typically those used in applied research (many private ones do not wish to publish such information for obvious reasons). For example, in Brazil – which shows good computing capacity with 88,175 TFlops – the vast majority is in the hands of the oil industry and only about 3,000 TFlops are used for basic research. Countries listed in the TOP500 invest in HPC from a few TFlops per million GDP (Belgium 5, Spain 7, Bulgaria 8), through countries investing in the order of hundreds (Italy 176, Japan 151, USA 138), to even thousands, as is the case in Finland with 1,478. For those countries where we were able to find data on the number of researchers, these range from a few Gflops per researcher (Belgium 19, Spain 24, Hungary 52) to close to 1,000 GFlops, i.e. 1 TFlop (USA 970, Italy 966), with Finland surpassing this barrier with 4,647. Note that, unlike what happens locally, countries with a certain degree of development invest every 3-4 years in supercomputing, so the data we are showing will soon be updated and there will be variations in the list. For example, this year a new supercomputer will come into operation in Spain(4), which, with an investment of some 150 million euros, will give Spain one of the most powerful supercomputers in Europe – and the world.
Country | Rpeak
(TFlops) |
Population
(millions) |
Res/1000 | GFlops/Res | Tflops/M US$ |
United States | 3.216.124 | 335 | 9.9 | 969.7 | 138.0 |
Canada | 71.911 | 39 | 8.8 | 209.5 | 40.0 |
Brazil | 88.175 | 216 | 1.1 | 371.1 | 51.9 |
AMERICA | 3.376.211 | 590 | |||
China | 1.132.071 | 1400 | 67.4 | ||
Japan | 815.667 | 124 | 10.0 | 657.8 | 151.0 |
South Korea | 128.264 | 52 | 16.6 | 148.6 | 71.3 |
Saudi Arabia | 98.982 | 35 | 141.4 | ||
Taiwan | 19.562 | 23 | 21.7 | ||
Singapore | 15.785 | 6 | 52.6 | ||
Thailand | 13.773 | 70 | 27.5 | ||
United Arab Emirates | 12.164 | 10 | 15.2 | ||
India | 12.082 | 1380 | 4.0 | ||
ASIA | 2.248.353 | 3100 | |||
Finland | 443.391 | 6 | 15.9 | 4647.7 | 1478.0 |
Italy | 370.262 | 59 | 6.5 | 965.5 | 176.3 |
Germany | 331.231 | 85 | 10.1 | 385.8 | 78.9 |
France | 251.166 | 65 | 11.4 | 339.0 | 83.7 |
Russia | 101.737 | 145 | 59.8 | ||
United Kingdom | 92.563 | 68 | 9.6 | 141.8 | 29.9 |
Netherlands | 56.740 | 18 | 10.6 | 297.4 | 56.7 |
Switzerland | 38.600 | 9 | 9.4 | 456.3 | 48.3 |
Sweden | 32.727 | 10 | 15.8 | 207.1 | 54.5 |
Ireland | 26.320 | 5 | 10.6 | 496.6 | 65.8 |
Luxembourg | 18.291 | 0.6 | 365.8 | ||
Poland | 17.099 | 38 | 7.6 | 59.2 | 28.5 |
Norway | 17.031 | 6 | 13.0 | 218.3 | 34.1 |
Czech Republic | 12.914 | 10 | 8.3 | 155.6 | 43.0 |
Spain | 10.296 | 47 | 7.4 | 29.6 | 7.4 |
Slovenia | 10.047 | 2 | 9.9 | 507.4 | 167.5 |
Austria | 6.809 | 9 | 11.6 | 65.2 | 13.6 |
Bulgaria | 5.942 | 6 | 8.5 | ||
Hungary | 4.669 | 10 | 9.0 | 51.9 | 23.3 |
Belgium | 3.094 | 12 | 13.6 | 19.0 | 5.2 |
EUROPA | 1.850.934 | 610.6 | |||
OTHER | |||||
Australia | 60.177 | 26 | 40.1 | ||
Morocco | 5.014 | 39 | 50.1 |
Table 1. HPC availability per researcher and relative to GDP in the TOP500 countries (includes HPC in industry).
The local reality is far from this data. Table 2 shows data from Argentina, Brazil, Chile and Mexico. In Chile, the availability of computing power is 2-3 times less per researcher than in countries with less computing power in the OECD and up to 100 times less than a researcher in the US. In Chile, our investment measured in TFlops per million US$ of GDP is 166 times less than in the US; with respect to European countries that invest less in HPC it is 9 times less, and with respect to the European average (including Finland) it is 80 times less, i.e. the difference is considerable. It is clear that we need to close this gap. An investment go about 5 million dollars in HPC infrastructure in the next 5 years would close this gap by a factor of almost 20 times our computational capacity. However, returning to the example of Spain, the supercomputer it will have this year will offer 23 times more computing power than at present and, therefore, we will only maintain our relative distance. If we do not invest, the dap will increase by at least 23 times and will end up being huge. Therefore, we do not only need a one-time investment, but we need to ensure a regular investment. Some neighbouring countries are already investing significantly in supercomputing. This is the case in Argentina, where they are investing 7 million dollars (2 million for the datacenter and 5 million to buy a new supercomputer), which will increase their current capacities by almost 40 times(5).
Country | Rpeak
(TFlops) |
Population (millions) | Res/1000 | GFlops/Res | Tflops/M US$ |
Brazil* | 3.000 | 216 | 1.1 | 12.6 | 1.8 |
Mexico | 2.200 | 130 | 1.2 | 14.1 | 1.8 |
Argentina | 400 | 45 | 1.2 | 7.4 | 0.8 |
Chile | 250 | 20 | 1.3 | 9.6 | 0.8 |
Table 2. HPC availability per researcher and relative to GDP in the region (*only HPC capacity in academia is considered in this table).
For the above reasons, we are working to convince the Chilean authorities that we must have greater funding and, more crucially, permanent state funding in HPC. In relation to this, on July 6 we signed a collaboration agreement between 44 institutions with the support of the Ministry of Science to work on the creation of the National Supercomputing Laboratory(6). The agreement recognised that supercomputers are a critical infrastructure for Chile’s development, that it is necessary to centralise the requirements/resources at the national level, obtain permanent funding from the State and create a new institutional framework to provide governance. In an unprecedented inter-institutional collaboration in Chile, the competition for HPC resources at the national level is eliminated ad the possibility of direct funding from the State is opened up without generating controversy.
Undoubtedly, supercomputing is a fundamental pillar for the development of any country, where increasing investment provides a strategic advantage, and in Latin America we should not be left behind.
By NLHPC
References
(1) Hyperion Research HPC Investments Bring High Returns
(5) https://www.hpcwire.com/2022/12/15/argentina-announces-new-supercomputer-for-national-science/
The post Towards a greater HPC capacity in Latin America first appeared on RISC2 Project.
]]>The post Costa Rica HPC School 2023 aimed at teaching the fundamental tools and methodologies in parallel programming first appeared on RISC2 Project.
]]>Building on the success of previous editions, the seventh installment of the Costa Rica High Performance Computing School (CRHPCS) aims at preparing students and researchers to introduce HPC tools in their workflows. A selected team of international experts taught sessions on shared-memory programming, distributed-memory programming, accelerator programming, and high performance computing. This edition had instructors Alessandro Marani and Nitin Shukla from CINECA, which greatly helped in bringing a vibrant environment to the sessions.
Bernd Mohr, from Jülich Supercomputing Centre, was the Keynote Speaker of this year’s edition of the event. A well-known figure in the HPC community at large, Bernd presented the talk Parallel Performance Analysis at Scale: From Single Node to one Million HPC Cores. In an amazing voyage through different architecture setups, Bernd highlighted the importance and challenges of performance analysis.
For Esteban Meneses, Costa Rica HPC School General Chair, the School is a key element in building a stronger and more connected HPC community in the region. This year, thanks to the RISC2 project, we were able to gather participants from Guatemala, El Salvador, and Colombia. Creating these ties is fundamental for later developing more complex initiatives. We aim at preparing future scientists that will develop groundbreaking computer applications that tackle the most pressing problems of our region.
The post Costa Rica HPC School 2023 aimed at teaching the fundamental tools and methodologies in parallel programming first appeared on RISC2 Project.
]]>