
Andrew J. YoungeSandia National Laboratories · Computer Science Research Institute
Andrew J. Younge
Ph.D Computer Science
Manager of Scalable Computer Architecture department at Sandia. Research in HPC architectures and operating systems.
About
70
Publications
63,411
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
2,518
Citations
Introduction
Andrew J. Younge is a Computer Scientist in the Scalable System Software department at Sandia National Laboratories. His research interests include high performance computing, virtualization, distributed systems, and energy efficient computing. The focus of his research is on improving the usability and efficiency of system software for supercomputing systems. Dr. Younge currently serves as the Lead PI for the Supercontainers project under the DOE Exascale Computing Project and is a key contributor to the Astra system, the world's first supercomputer based on the Arm processor deployed under Sandia's Vanguard program. He received his PhD in computer science from Indiana University in 2016 and his BS and MS in computer science from the RIT in 2008 and 2010, respectively.
Publications
Publications (70)
The Cloud computing emerges as a new computing paradigm which aims to provide reliable, customized and QoS guaranteed dynamic computing environments for end-users. In this paper, we study the Cloud computing paradigm from various aspects, such as denitions, distinct features, and enabling technologies. This paper brings an introductional review on...
As Cloud computing emerges as a dominant paradigm in distributed systems, it is important to fully understand the underlying technologies that make Clouds possible. One technology, and perhaps the most important, is virtualization. Recently virtualization, through the use of hyper visors, has become widely used and well understood by many. However,...
Arm processors have been explored in HPC for several years, however there has not yet been a demonstration of viability for supporting large-scale production workloads. In this paper, we offer a retrospective on the process of bringing up Astra, the first Petascale supercomputer based on 64-bit Arm processors, and validating its ability to run prod...
Astra, deployed in 2018, was the first petascale supercomputer to utilize processors based on the ARM instruction set. The system was also the first under Sandia's Vanguard program which seeks to provide an evaluation vehicle for novel technologies that with refinement could be utilized in demanding, large‐scale HPC environments. In addition to ARM...
HPC centers face increasing demand for software flexibility, and there is growing consensus that Linux containers are a promising solution. However, existing container build solutions require root privileges and cannot be built directly on HPC resources. This limitation is compounded as supercomputer diversity expands and HPC architectures become m...
Several recent workshops conducted by the DOE Advanced Scientific Computing Research program have established the fact that the complexity of developing applications and executing them on high-performance computing (HPC) systems is rising at a rate which will make it nearly impossible to continue to achieve higher levels of performance and scalabil...
Astra, deployed in 2018, was the first petascale supercomputer to utilize processors based on Arm’s instruction set. The system was also the first under Sandia’s Vanguard program which seeks to provide an evaluation vehicle for novel technologies that with refinement could be utilized in demanding, large-scale HPC environments. While the adoption o...
Trusting simulation output is crucial for Sandia’s mission objectives. We rely on these simulations to perform our high-consequence mission tasks given national treaty obligations. Other science and modeling applications, while they may have high-consequence results, still require the strongest levels of trust to enable using the result as the foun...
Containers offer a broad array of benefits, including a consistent lightweight runtime environment through OS-level virtualization, as well as low overhead to maintain and scale applications with high efficiency. Moreover, containers are known to package and deploy applications consistently across varying infrastructures. Container orchestrators ma...
Containerized computing is quickly changing the landscape for the development and deployment of many HPC applications. Containers are able to lower the barrier of entry for emerging workloads to leverage supercomputing resources. However, containers are no silver bullet for deploying HPC software and there are several challenges ahead in which the...
A class of efficient processing methods known as Streaming Analytics is emerging within
Distributed Systems to analyze the contents of high velocity, high volume data streams in real
time. Streaming Analytics can potentially solve a variety of computationally hard problems.
However, realizing these solutions depends on the ability to feed high-thro...
Power measurement capabilities are becoming commonplace on large scale HPC system deployments. There exist several different approaches to providing power measurements that are used today, primarily in-band and out-of-band measurements. Both of these fundamental techniques can be augmented with application-level profiling and the combination of dif...
Trusting simulation output is crucial for Sandia's mission objectives. We rely on these simulations to perform our high-consequence mission tasks given national treaty obligations. Other science and modeling applications, while they may have high-consequence results, still require the strongest levels of trust to enable using the result as the foun...
Large-scale HPC systems increasingly incorporate sophisticated power management control mechanisms. While these mechanisms are potentially useful for performing energy and/or power-aware job scheduling and resource management (EPA JSRM), greater understanding of their operation and performance impact on real-world applications is required before th...
Advanced power measurement capabilities are be- coming available on large scale High Performance Computing (HPC) deployments. There exist several approaches to providing power measurements today, primarily through in-band (e.g. RAPL) and out-of-band measurements (e.g. power meters). Both types of measurement can be augmented with application-level...
The Observatory on Social Media (OSoMe) provides a Terabyte-scale historical and ongoing collection of approximately 70 billion public tweets.
This covers Streaming workshops held, IoTCloud for cloud control of robots, SPIDAL project, HPC-ABDS, WebPlotviz visualization and Stock Market data, Scientific paper impact analysis for XSEDE
This poster covers the Harp HPC Hadoop plugin, RaPyDLI deep learning system, Virtual Clusters on XSEDE Comet system, Cloudmesh to defer Ansible Big data applications, Big Data Ogres and Diamonds to converge HPC and Big Data, Performance of Flink on machine learning
This poster introduces all of DSC projects below and covers 1) 3) 4) 5) 1) Digital Science Center Facilities 2) RaPyDLI Deep Learning Environment 3) SPIDAL Scalable Data Analytics Library and applications including Bioinformatics and Polar Remote Sensing Data Analysis 4) MIDAS Big Data Software; Harp for HPC-ABDS 5) Big Data Ogres Classification an...
With the advent of virtualization and Infrastructure-as-a-Service (IaaS), the broader scientific computing community is considering the use of clouds for their scientific computing needs. This is due to the relative scalability, ease of use, advanced user environment customization abilities, and the many novel computing paradigms available for data...
The study of social phenomena is becoming increasingly reliant on big data from online social networks. Broad access to social media data, however, requires software development skills that not all researchers possess. Here we present the IUNI Observatory on Social Media , an open analytics platform designed to facilitate computational social scien...
While distributed memory systems have shaped the field of distributed systems for decades, the demand for many-core shared memory resources is increasing. Symmetric Multiprocessor Systems (SMPs) have become increasingly important recently among a wide array of disciplines, ranging
from Bioinformatics to astrophysics, and beyond. With the increase i...
The study of social phenomena is becoming increasingly reliant on big data from online social networks. Broad access to social media data, however, requires software development skills that not all researchers possess. Here we present the IUNI Observatory on Social Media, an open analytics platform designed to facilitate computational social scienc...
The study of social phenomena is becoming increasingly reliant on big data from online social networks. Broad access to social media data, however, requires software development skills that not all researchers possess. Here we present the IUNI Observatory on Social Media, an open analytics platform designed to facilitate computational social scienc...
The study of social phenomena is becoming increasingly reliant on big data from online social networks. Broad access to social media data, however, requires software development skills that not all researchers possess. Here we present the IUNI Observatory on Social Media, an open analytics platform designed to facilitate computational social scienc...
Knowledge of the genome-wide rate and spectrum of mutations is necessary to understand the origin of disease and the genetic variation driving all evolutionary processes. Here, we provide a genome-wide analysis of the rate and spectrum of mutations obtained in two Daphnia pulex genotypes via separate mutation-accumulation (MA) experiments. Unlike m...
Cloud Infrastructure-as-a-Service paradigms have recently shown their utility for a vast array of computational problems, ranging from advanced web service architectures to high throughput computing. However, many scientific computing applications have been slow to adapt to virtualized cloud frameworks. This is due to performance impacts of virtual...
The intensive research activity in analysis of social media and micro-blogging data in recent years suggests the necessity and great potential of platforms that can efficiently store, query, analyze, and visualize social media data. To support these “social media observatories” effectively, a storage platform must satisfy special requirements for l...
The role of network structure in intergroup differentiation-the bipolarization of stereotypes that are defensive (ingroup-positive/outgroup-negative) and non-defensive (outgroup-positive/ingroup-negative)-was investigated using a Dynamic Social Impact Theory (DSIT) framework. Three computer-mediated laboratory social network experiments were pooled...
As more scientific workloads are moved into the cloud, the need for high performance accelerators increases. Accelerators such as GPUs offer improvements in both performance and power efficiency over traditional multi-core processors, however, their use in the cloud has been limited. Today, several common hypervisors support GPU passthrough, but th...
With the advent of virtualization and Infrastructure-as-a-Service (IaaS), the broader scientific computing community is considering the use of clouds for their technical computing needs. This is due to the relative scalability, ease of use, advanced user environment customization abilities clouds provide, as well as many novel computing paradigms a...
With the advent of virtualization and Infrastructure-as-a-Service (IaaS), the broader scientific computing community is considering the use of clouds frothier scientific computing needs. This is due to the relative scalability, ease of use, advanced user environment customization abilities, and the many novel computing paradigms available for data-...
Through the development of advanced middleware, Grid computing has evolved to a mature technology that scientists and researchers can leverage to gain knowledge that was previously unobtainable in a wide variety of scientific
Future Grid (FG) is an experimental, high-performance test bed that supports HPC, cloud and grid computing experiments for both application and computer scientist. Future Grid includes the use of virtualization technology to allow the support of a wide range of operating systems in order to include a test bed for various cloud computing infrastruct...
In this paper, we briefly outline the current design of a generic image management service for FutureGrid. The service is intended to generate, store, and verify images while interfacing with different localized cloud IaaS image. Additionally, we will also use the service to generate images for traditional bare-metal deployments.
As today's fast paced environment continually encourages poor dietary habits and a lack of exercise, there is a growing need to properly monitor and control weight gain. With the advent of pervasive and ubiquitous computing, there are new opportunities to help promote personal wellness that was previously unobtainable. This work describes the novel...
With the advent of Cloud computing, a wide variety of In-frastructure as a Service models have grown to provide users with one of the greatest benefits of Clouds: a customized system environment. These services, while extremely use-ful, often suffer from their ability to interoperate and com-municate across administratively separate domains. Within...
FutureGrid provides novel computing capabilities that enable reproducible experiments while simultaneously supporting dynamic provisioning. This paper describes the FutureGrid experiment management framework to create and execute large scale scientific experiments for researchers around the globe. The experiments executed are performed by the vario...
The notion of Cloud computing has not only reshaped the field of distributed systems but also fundamentally changed how businesses utilize computing today. While Cloud computing provides many advanced features, it still has some shortcomings such as the relatively high operating cost for both public and private Clouds. The area of Green computing i...
Using networked computers in laboratory experiments to investigate group dynamics currently requires the creation of special program application software. Design considerations for a useable graphical user interface (GUI) in such software are discussed in this paper. We describe SOCIALSENSE communication software created to experimentally test the...
As Cloud computing emerges as the dominant paradigm in distributed systems, it's important to fully under-stand the underlying technologies that make clouds possible. One technology, and perhaps the most important, is virtualization. Recently virtualization through the use of hypervisors has become widespread and well understood by many. However, t...
With the advent of Cloud computing, large-scale virtualized compute and data centers are becoming common in the computing industry. These distributed systems leverage commodity server hardware in mass quantity, similar in theory to many of the fastest Supercomputers in existence today. However these systems can consume a cities worth of power just...
Abstract—In recent years the power,of Grid computing,has grown,exponentially,through,the development,of advanced,mid- dleware systems. While usage has increased, the penetration of Grid computing,in the scientific community,has been less than expected,by some. This is due,to a steep learning,curve,and high,entry barrier that limit the use of Grid c...
High density blade servers are a popular technology for data centers, however, the heat dissipation density of data centers increases exponentially. There is strong evidence to support that high temperatures of such data centers will lead to higher hardware failure rates and thus an increase in maintenance costs. Improperly designed or operated dat...
In this paper, we describe a service oriented architecture and Grid abstraction framework that allows us to access Grids through JavaScript. Obviously, such a framework integrates well with other Web 2.0 technologies. The framework consists of two parts. A client Application Programming Interface (API) to access the Grid via JavaScript and a mediat...
Intelligent agents can be used as a valuable asset when deployed within a wireless sensor network. Wireless sensor networks have a number of limitations, which need to be addressed before creating successful agents. These limitations include issues with energy conservation, mitigating resource limitations and wireless link availability to name a fe...
Cloud computing has and will continue to emerge as a fundamental paradigm shift within Distributed Systems. While the features of Clouds are well known and widespread, the ability for users to specifically define and customize their environment is paramount to the advancement of Clouds. As such, there is an ever-growing need to leverage the ability...