
Dmitry NikitenkoLomonosov Moscow State University | MSU · Research Computing Center
Dmitry Nikitenko
PhD
About
24
Publications
2,428
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
295
Citations
Introduction
Publications
Publications (24)
There is a variety of known HPC ratings nowadays which represent machine capability for solving a fixed problem, based on a certain algorithm, but these ratings represent a top of the iceberg, and as a rule, one can’t compare application tuning features even for the selected system, and the details of system architecture are not usually described p...
In order to ensure high performance of a large supercomputer facility, its management must collect and analyze information on many different aspects of this complex operation in a timely manner. One of the most important directions in this area is to analyze the supercomputer workload. In particular, it is necessary to detect the subject areas bein...
Managing and administering an HPC center is a real challenge. The Supercomputing Center at Moscow State University is the largest in Russia, with petascale machines running in the interest of thousands of users engaged in hundreds of research projects. Since its early stages of development, the Octoshell system has been instrumental in mastering th...
The main goal of this work is to demonstrate that the development of data-intensive applications for vector systems is not only important and interesting, but is also very possible. In this paper we describe possible implementations of two fundamental graph-processing algorithms for an NEC SX-ACE vector computer: the Bellman-Ford algorithm for sing...
The resource utilization analysis of HPC systems can be performed in different ways. The method of analysis is selected depending primarily on the original focus of research. It can be a particular application and/or a series of application run analyses, a selected partition or a whole supercomputer system utilization study, a research on peculiari...
It is a common knowledge that the increasingly growing capabilities of HPC systems are always limited by a number of efficiency related issues. The reasons can be very different: hardware failures, incorrect job scheduling, peculiarities of algorithm, chosen programming technology specifics, etc. Most of these issues can be detected after precise a...
The availability of high performance computing resources enables us to perform very large numerical simulations and in this way to tackle challenging real life problems. At the same time, in order to efficiently utilize the computational power at our disposal, the ever growing complexity of the computer architecture poses high demands on the algori...
The efficiency of computing resources utilization by user applications can be analyzed in various ways. The JobDigest approach based on system monitoring was developed in Moscow State University and is currently used in everyday practice of the largest Russian supercomputing center of Moscow State University. The approach features application behav...
The most productive and powerful supercomputers always arouse great interest being flagships of the HPC machines fleet. These giants always require new forms of presentation because of innovations in design. This also leads to evaluation and development of new ways of comparison and supercomputer ranking. There is a variety of known regional and wo...
The problem of effective resource utilization is very challenging nowadays, especially for HPC centers running top-level supercomputing facilities with high energy consumption and significant number of workgroups. The weakness of many system monitoring based approaches to efficiency study is the basic orientation on professionals and analysis of sp...
Efficient use and high output of any supercomputer depends on a great number of factors. The problem of controlling granted resource utilization is one of those, and becomes especially noticeable in conditions of concurrent work of many user projects. It is important to provide users with detailed information on peculiarities of their executed jobs...
The efficiency of most supercomputer applications is extremely low. At the same time, the user rarely even suspects that their applications may be wasting computing resources. Software tools need to be developed to help detect inefficient applications and report them to the users. We suggest an algorithm for detecting anomalies in the supercomputer...
The effective mastering of extremely parallel HPC system is impossible without deep understanding of all internal processes and behavior of the whole diversity of the components: computing processors and nodes, memory usage, interconnect, storage, whole software stack, cooling and so forth in detail. There are numerous visualization tools that prov...
Managing and administering of large-scale HPC centers is a complicated problem. Using a number of independent tools for resolving its seemingly independent sub problems can become a bottleneck with rapidly increasing scale of systems, number of hardware and software components, variety of user applications and types of licenses, number of users and...
In this article we describe the Octotron project intended to ensure reliability and sustainability of a supercomputer. Octotron is based on a formal model of computing system that describes system components and their interconnections in graph form. The model determines relations between data describing current supercomputer state (monitoring data)...
The paper introduces techniques for solving various large-scale graph problems on hybrid architectures. The proposed approach is illustrated on the computation of minimum spanning tree and shortest paths. We provide a precise mathematical description accompanied by the information structure of required algorithms. Efficient parallel implementations...
To maximise the scientific output of a high-performance computing system, different stakeholders pursue different strategies. While individual application developers are trying to shorten the time to solution by optimising their codes, system administrators are tuning the configuration of the overall system to increase its throughput. Yet, the comp...
Иногда высказывается мнение, что, хотя суперкомпьютеры и становятся все мощнее, они не всегда ориентированы на потребности конечного пользователя и используются неэффективно. Опыт эксплуатации суперкомпьютера «Ломоносов» показывает, что это не так.
With the scale of supercomputing systems and applications growing fast, the difficulty of developing performance efficient applications also grows rapidly. The reason for this is an extensive number of factors that potentially influence the application performance. Hardware and software specifics of the supercomputer, peculiarities of the applicati...