
Andrzej Goscinski- PhD DSc Professor
- Professor (Full) at Deakin University
Andrzej Goscinski
- PhD DSc Professor
- Professor (Full) at Deakin University
About
256
Publications
38,297
Reads
How we measure 'reads'
A 'read' is counted each time someone views a publication summary (such as the title, abstract, and list of authors), clicks on a figure, or views or downloads the full-text. Learn more
2,905
Citations
Introduction
Current institution
Additional affiliations
January 1992 - present
Publications
Publications (256)
In the last decade there has been a strong move towards mobile computing and the proliferation of the IoT (Internet of Things). A huge number of devices have been connected to the Internet and created zettabytes of data items. To extract value from such massive data volumes, processing power offered by cloud computing is often utilized. However, st...
Service-Oriented Computing (SOC) and SOC systems have been invented, studied, and developed in response to the problems generated by heterogeneity and poor latency, and a need for modularization and standardization. Wide application of SOC, practically in all areas of human endeavours, interconnectivity provided by wireless networks, huge volumes a...
The concept of Social Internet of Things (SIoT) has emerged from the integration of social networking into the core of the Internet of Things (IoT). It envisions IoT objects and devices to have social interactions with each other autonomously, cooperate with other agents, and exchange information with human users and surrounding computing devices....
The overall process of discovering hydrocarbon traps, starting with geological exploration through to Seismic Data Processing (SDP) is very expensive and time consuming. In the real-world, the oil and gas production relies on how soon seismic data is computationally processed. The ability for an oil and gas company to perform seismic computation at...
By moving data computation and service supply from the cloud to the edge, edge computing has become a promising solution to address the limitations of cloud computing in supporting delay-sensitive and context-aware services in the Internet of Things (IoT) era. Instead of performing data storage and computing in a cluster of clouds, edge computing e...
SCADA systems allow users to monitor and/or control physical devices, processes, and events remotely and in real-time. As these systems are critical to industrial processes, they are often run on highly reliable and dedicated hardware. Moving these SCADA systems to an Infrastructure as a Service (IaaS) cloud allows for: cheaper deployments, system...
SCADA (Supervisory Control And Data Acquisition) systems are instrumental to a wide range of mission-critical industrial systems, such as gas pipelines, water control facilities and industrial plants.
SCADA systems allow a user to monitor (using sensors) and control (using switches) an industrial system remotely. Larger SCADA systems can support se...
Stream-oriented applications account for one of the major types of today's computing practices. They deal with high-speed data streams. When it comes to stream-oriented distributed systems, some challenges arise: for one thing, processing capability of a single compute node might become the bottleneck of the entire system. For another, transmitting...
SCADA systems allow users to monitor and/or control physical devices, processes and events remotely. As these systems are critical to industrial processes, they are often run on highly reliable and dedicated hardware. Moving these SCADA systems to an Infrastructure as a Service (IaaS) cloud, allows for: cheaper deployments, system redundancy suppor...
Our undergraduate curriculum focuses on preparing students to be critical thinkers, problem solvers, and knowledgeable, responsible information technology (IT) professionals. Students often question the relevance of core subjects which are designed to introduce and develop these attributes; such units regularly perform poorly in student evaluation...
Since the development of the computer, user orientated innovations such as graphical operating systems, mice, and mobile devices have made computing ubiquitous in modern society. The cloud is the next step in this process. Through the cloud, computing has undergone co modification and has been made available as a utility. However, in comparison to...
In current digital era according to (as far) massive progress and development of internet and online world technologies such as big and powerful data servers we face huge volume of information and data day by day from many different resources and services which was not available to human kind just a few decades ago. This data comes from available d...
Cloud and service computing has started to change the way research in science, in particular biology and medicine, is being carried out. Researchers that have taken advantage of this technology (making use of public and private cloud compute resources) can process large amounts of data (big data) and speed up discovery. However, this requires resea...
While High Performance Computing clouds allow researchers to process large amounts of genomic data, complex resource and software configuration tasks must be carried out beforehand. The current trend exposes applications and data as services, simplifying access to clouds. This paper examines commonly used cloud-based genomic analysis services, intr...
By processing big data using clouds, scientific researchers can achieve remarkable outcomes. However, these non-computing specialists do not have the computing knowledge and skills to deal with big data, build HPC applications, and execute them on clouds. Non-computing specialists also face a major problem with accessing HPC and cloud resources thr...
This paper introduces ePortfolio and its use of managing learning evidence in an eLearning envi-ronment. The paper uses the outcomes of a recent project sponsored by Deakin University, Aus-tralia and examples from IT teaching practice based on CloudDeakin to explore how ePortfolios add value to personalized and reflective models of eLearning. ePort...
Cloud-based service computing has started to change the way how research in science, in particular biology, medicine, and engineering, is being carried out. Researchers in the area of mammalian genomics have taken advantage of cloud computing technology to cost-effectively process large amounts of data and speed up discovery. Mammalian genomics is...
Supervisory Control and Data Acquisition (SCADA) systems control and monitor industrial and critical infrastructure functions, such as electricity, gas, water, waste, railway, and traffic. Recent attacks on SCADA systems highlight the need for stronger SCADA security. Thus, sharing SCADA traffic data has become a vital requirement in SCADA systems...
Although Australian students spend three or more years studying they can seem quite unaware of any of the expected learning outcomes of their course. They are often single unit focused, paying most attention to individual assessment items thus not developing a holistic view of their course. This paper presents a theoretical framework to support sta...
High Performance Computing (HPC) clouds have started to change the way how research in science, in particular medicine and genomics (bioinformatics) is being carried out. Researchers who have taken advantage of this technology can process larger amounts of data and speed up scientific discovery. However, most HPC clouds are provided at an Infrastru...
Plug-in Electric Vehicles (PEVs) provide new opportunities to reduce fuel consumption and exhaust emission. PEVs need to draw and store energy from an electrical grid to supply propulsive energy for the vehicle. As a result, it is important to know when PEVs batteries are available for charging and discharging. Furthermore, battery energy managemen...
Clouds have provided on-demand, scalable and affordable High Performance Computing (HPC) resources to discipline (e.g., Biology, Medicine, Chemistry) scientists. However, the steep learning curve of preparing a HPC cloud and deploying HPC applications has hindered many scientists to achieve innovative discoveries for which HPC resources must be rel...
This paper firstly introduces the concept of constructivist learning which advocates that students actively construct knowledge themselves with teachers’ assistance. Based on the six important elements of constructivist learning and teaching planning approach, detailed examples of designing the six constructivist elements of situation, groupings, b...
This paper firstly explains the importance of assessment management, then introduces two assessment tools currently used in the School of Information Technology at Deakin University. A comparison of assignment marking was conducted after collecting test data from three sets of assignments. The importance of providing detailed marking guides and per...
The system biology approach has been applied to gene expression data to identify molecular mechanisms of isolated biological systems. Adapting this methodology to incorporate connections between different biological systems and extensive repositories of data can help find novel gene interactions. The devised distributed workflow uses normalized mic...
This paper discusses a few issues related to teaching improvement that are commonly found in tertiary education, such as curriculum development, student engagement, and ethical considerations. Scholars re- search on resolving these issues are investigated. Corresponding approaches to improving teaching of a year one information technology unit are...
VMD and NAMD are two major molecular dynamics simulation software packages, which can work together for mining structural information of bio-molecules. Carrying out such molecular dynamics simulations can help researchers to understand the roles and functions of various bio-molecules in life science research. Recently, clouds have provided HPC clus...
High Performance Computing (HPC) clusters have enabled Molecular Dynamics (MD) simulations of biological systems for a sufficiently large complexity and long simulation period. VMD and NAMD are two major MD simulations software packages, which can work together for mining structural information of bio-molecules. However, HPC clusters are in many ca...
Virtualization is a rapidly evolving technology that can be used to provide a range of benefits to computing systems, including improved resource utilization, software portability, and reliability. Virtualization also has the potential to enhance security ...
The increasing significance of Distributed Computing becomes more and more crucial with the prevail of technological advances that make Global Computing a reality in modern world. Indeed, it is hard to imagine some application or computational activity and process that falls outside Distributed Computing. With the large advent of distributed system...
Cloud computing is the most recent realisation of computing as a utility. Recently, fields with substantial computational requirements, e.g., biology, are turning to clouds for cheap, on-demand provisioning of resources. Of interest to this paper is the execution of compute intensive applications on hybrid clouds. If application requirements exceed...
The cost and time of deploying HPC applications on clouds is a problem. Instead of conducting their research discipline specialists are forced to carry out activities for application deployment, publication and ease of access. In response, a new approach for HPC application deployment and access in clouds is proposed. The major innovations are a ne...
Seismic data gathered from the Hydrocarbon Exploration and Discovery Operation is essential to identify possible hydrocarbon existence in a geologically surveyed area. However, the discovery operation takes a long time to be completed and computational processing of the acquired data is often delayed. Hydrocarbon exploration may end up needlessly c...
Microarrays and more recently RNA sequencing has led to an increase in available gene expression data. How to manage and store this data is becoming a key issue. In response we have developed EXP-PAC, a web based software package for storage, management and analysis of gene expression and sequence data. Unique to this package is SQL based querying...
Recent developments in sensor networks and cloud computing saw the emergence of a new platform called sensor-clouds. While the proposition of such a platform is to virtualise the management of physical sensor devices, we are seeing novel applications been created based on a new class of social sensors. Social sensors are effectively a human-device...
Expressed Sequence Tags (ESTs) are short DNA sequences generated by sequencing the transcribed cDNAs coming from a gene expression. They can provide significant functional, structural and evolutionary information and thus are a primary resource for gene discovery. EST annotation basically refers to the analysis of unknown ESTs that can be performed...
Recent developments in sensor networks and cloud computing saw the emergence of a new platform called sensor-clouds. While the proposition of such a platform is to virtualise the management of physical sensor devices, we foresee novel applications being created based on a new class of social sensors. Social sensors are effectively a human-device co...
Nowadays cloud computing has become a major trend that enterprises and research organizations are pursuing with increasing zest. A potentially important application area for clouds is data analytics. In our previous publication, we introduced a novel cloud infrastructure, the CloudMiner, which facilitates data mining on massive scientific data. By...
In many businesses, including hydrocarbon industries, reducing cost is of high priority. Although hydrocarbon industries appear able to afford the expensive computing infrastructure and software packages used to process seismic data in the search for hydrocarbon traps, it is always imperative to find ways to minimize cost. Seismic processing costs...
The challenge of comparing two or more genomes that have undergone recombination and substantial amounts of segmental loss and gain has recently been addressed for small numbers of genomes. However, datasets of hundreds of genomes are now common and their sizes will only increase in the future. Multiple sequence alignment of hundreds of genomes rem...
Gene Expression Comparative Analysis allows bio-informatics researchers to discover the conserved or specific functional regulation of genes. This is achieved through comparisons between quantitative gene expression measurements obtained in different species on different platforms to address a particular biological system. Comparisons are made more...
Business, scientific and engineering experiments, medical studies, and governments generate huge amount of information. The problem is how to extract knowledge from all this information. Data mining provides means for at least a partial solution to this problem. However, it would be too expensive to all these areas of human activity and companies t...
IntroductionRelated WorkRVWS DesignCluster as a Service: The Logical DesignProof of ConceptFuture Research DirectionsConclusion
References
The primary purpose of this book is to capture the state-of-the-art in Cloud Computing technologies and applications. The book will also aim to identify potential research directions and technologies that will facilitate creation a global market-place of cloud computing services supporting scientific, industrial, business, and consumer applications...
The general idea of cloud computing is offering computational resources as scalable, on demand services accessible over the internet. However, this new realisation of on demand computing is hindered by the amount of user involvement. Currently, high level abstractions in cloud computing only exist in the form of services. In particular, there is no...
Cloud computing is an emerging paradigm where computing resources are offered over the Internet as scalable, on-demand (Web) services. While cloud vendors have concentrated their efforts on the improvement of performance, resource consumption and scalability, other cloud characteristics have been neglected. On the one hand cloud service providers f...
While the emergence of clouds had lead to a significant paradigm shift in business and research, cloud computing is still in its infancy. Specifically, there is no effective publication and discovery service nor are cloud services easy to use. This paper presents a new technology for offering ease of discovery, selection and use of clusters hosted...
When it comes to grid and cloud computing, there is a lot of debate over their relations to each other. A common feature is that grids and clouds are attempts at utility computing. However, how they realize utility computing is different. The purpose of this paper is to characterize and present a side by side comparison of grid and cloud computing...
The goal of this paper is to present the experiences gained over 15 years of research into the design and development of a
services-based distributed operating system. The lessons learnt over this period, we hope, will be of value to researchers
involved in the design and development of operating systems that wish to harness the collective resource...
Gene Expression Comparative Analysis allows bio-informatics researchers to discover the functional regulation of genes. This
is achieved through comparisons between data-sets representing the quantities of substances in a biological system. Unnatural
variations can be introduced during the data collection and digitization process so normalization a...
While the emergence of cloud computing has made it possible to rent information technology infrastructures on demand, it has also created new security challenges. The primary security concern is trusting data (or resources in general) on another organization's system. This document seeks to examine the current state of security in cloud computing a...
Clouds refer to computational resources (in particular, clusters) that are accessible as scalable, on demand, pay-as-you-go services provided in the Internet. However, clouds are in their infancy and lack a high level abstraction. Specifically, there is no effective discovery and selection service for clusters and offer little to no ease of use for...
This paper presents cloud's higher layer abstraction and support for users in the form of cluster as a service.
A variety of research problems exist that require considerable time and computational resources to solve. Attempting to solve these problems produces long-running applications that require a reliable and trustworthy system upon which they can be executed. Cluster systems provide an excellent environment upon which to run these applications because...
With the emergence of cloud computing, the need for flexible and detailed publication and selection of services that expose cloud resources is greatly stressed. While dynamic attributes have improved the publication and selection of resources in distributed systems, the use of dynamic attributes is yet to be tried in Web services: a key element tha...
An autonomic computing system is organized into building blocks that can be composed together to form a self-managing system.
Architecturally, this matches service-based computing systems, which are the outcomes of the most recent effort to provide
interoperability and usability through the use of services. Autonomic computing is an attractive info...
Cluster systems provide an excellent environment to run computation hungry applications. However, due to being created using commodity components they are prone to failures. To overcome these failures we propose to use rollback-recovery, which consists of the checkpointing and recovery facilities. Checkpointing facilities have been the focus of man...
We assert that companies can make more money and research institutions can improve their performance if inexpensive clusters and enterprise grids are exploited. In this paper, we have demonstrated that our claim is valid by showing the study of how programming environments, tools and middleware could be used for the execution of parallel and sequen...
Multi agent systems, Grid technology, Semantic Web, Autonomic Com-puting, and Web Intelligence paradigms are modern approaches in information technologies, which we put together in our research effort described in this pa-per to create a new-generation infrastructure called the Autonomic Wisdom Grid (AWG) with the mission to maintain, share, discov...
A common architecture in today's development of distributed systems is the service-oriented architecture (SOA) implemented using Web services. Until recently, it was difficult to build a SOA based grid/distributed system using Web services due to the inability to learn the state of services. The state of a Web service could only be accessed through...
Dynamic deployment of Web services is a term used frequently when describing the selection and deployment of a service to
a grid host. Although current grid systems (such as Globus) provide dynamic deployment, the requirements of the service being
deployed are not considered. Therefore truly dynamic deployment cannot be achieved as the services dep...
The issue of under-estimated length of jobs (parallel applications) on backfill-based scheduling is ignored in the current literature because users want to avoid their jobs to be killed when the requested time expires. Therefore, users prefer to over-estimate the length of their jobs. This paper shows the impact of underestimated length of jobs on...
Computers of a non-dedicated cluster are often idle (users attend meetings, have lunch or coffee breaks) or lightly loaded (users carry out simple computations to support problem solving activities). These underutilised computers can be employed to execute parallel applications. Thus, these computers can be shared by parallel and sequential applica...
Using the volatile memory of other computers within a system to store checkpoints is an alternative to the traditional approach of using stable storage. The objective of this study is to develop a storage mechanism using at-least-k delivery semantics. This semantics allows data to be saved to a minimum number of computers simultaneously using group...
Dedicated clusters are becoming commonly used for high performance parallel processing. Computers of a non-dedicated cluster are often idle or lightly loaded. These under utilised computers can be employed to execute parallel applications. Thus, they have to be shared by parallel and sequential applications, which could lead to the improvement of t...
Parallel transfer of large amounts of data is attractive as (i) the data can be received in less time and (ii) is more resilient to connection failures. This is often used in e-Science under the field of data sharing. Data from one research project should be easily assessable to others regardless of the size of the data thus parallelism is used to...
With the recent innovations in stateful web services, they are now being used to support the construction of distributed systems using software as a service. While the state of web services is preserved, the state is still hidden from clients thus searches for both functionality and state remains a two step process. Proposed in this report is the R...
The introduction of Web services into grids helped to address their two main obstacles to be embraced by business and industry, heterogeneity and useability. However, many problems are still open, e.g., grid reconfiguration, reliability and computing optimization. We argue here that a mechanism that could help solving these problems is Web Service...
Although individual PCs of a cluster are used by their owners to run sequential applications (local jobs), the cluster as a whole or its subset can also be employed to run parallel applications (cluster jobs) even during working hours. This implies that these computers have to be shared by parallel and sequential applications, which could lead to t...
This research aims at improving our understanding of backfilling job scheduling algorithms. The most frequently used algorithm, EASY-backfilling, was selected for a performance evaluation by scheduling static workloads of parallel jobs on a computer cluster. To achieve the aim, we have developed a batch job scheduler for Linux clusters, implemented...
Mobile device integration in grid environments is a challenge for many researchers. Due to the transient nature of mobile devices, service management is a critical, but often overlooked area of research. We propose a distributed broker responsible for the autonomic management of grid services. The broker provides self discovery and negotiation, sel...
The increasing popularity of a Cluster of Workstations (COW) for the execution of parallel applications can be attributed
to its impressive price to performance ratio. Unfortunately, currently available software to manage the execution of parallel
applications on COWs do not provide satisfactory levels of performance, nor do they provide the applic...
An enterprise has not only a single cluster but a set of geographically distributed clusters – they could be used to form
an enterprise grid. In this paper we show based on our case study that enterprise grids could be efficiently used as parallel
computers to carry out high-performance computing.
In a large-scale wide-area system such as the Grid, trust is a prime concern. The current generation of grid information services
lack the ability to determine how trustworthy a particular grid service provider or grid customer is likely to be. In this
paper, we propose a grid information service with reputation management facility and its underlyi...
Parallel computing is strongly influenced by the challenges of distributed systems, such as a need for a Single System Image,
resource sharing and allocation, failures and a need for fault tolerance, long latencies, network partition, disconnected
operation, demands of users wishing to solve more computationally and communication demanding problems...
With the rising demand for multimedia content and updates on the Internet, parallelism is adopted to both increase performance and to relieve the strain on the system hosting the desired content. However, parallelism is not a straight forward task and does incur issues of its own, mainly task coordination and resource discovery. Resources include t...
In this paper, we have demonstrated how the existing programming environments, tools and middleware could be used for the
study of execution performance of parallel and sequential applications on a non-dedicated cluster. A set of parallel and sequential
benchmark applications selected for and used in the experiments were characterized, and experime...
Nondedicated clusters are currently at the forefront of the development of high performance computing systems. These clusters
are relatively intolerant of hardware failures and cannot manage dynamic cluster membership efficiently. This report presents
the logical design of an innovative self discovery service that provides for automated cluster man...
Cluster systems are becoming more prevalent in today’s computer society and users are beginning to request that these systems
be reliable. Currently, most clusters have been designed to provide high performance at the cost of providing little to no
reliability. To combat this, this report looks at how a recovery facility, based on either a centrali...
The move towards web services in Grid computing requires mechanisms for services to maintain state. This is introduced by
the Web Services Resource Framework which provides a basis for web services to access stateful resources. While this allows
access to stateful resources, the web services themselves are not stateful. Currently, Grids require a l...
Information Technology (IT) changes very quickly and influences business, industry and the public in an enormous manner. Outsourcing of IT jobs to cheaper overseas labor and globalization of IT companies become a common practice. Graduates of IT university courses must be well prepared to address the needs and expectations of business, industry and...
Currently, coordinated scheduling of multiple parallel applications across computers has been considered as the critical factor
to achieve high execution performance. We claim in this report that the performance and costs of the execution of parallel
applications could be improved if not only dedicated clusters but also non-dedicated clusters were...
A major problem for a grid user is the discovery of currently available services. With large number of services, it is beneficial for a user to be able to discover the services that most closely match their requirements. This report shows how to extend some concepts of UDDI such that they are suitable for dynamic parameter based discovery of grid s...
Fast and reliable parallel algorithms for the basic problems of numerical mathematics and their effective implementation in easy-to-use portable software components are crucial for computational solution of scientific and engineering problems. Since in all domains of high performance computing, parallel execution routinely is considered as one of t...