Conference Paper

Scheduling in Bag-of-Task Grids: The PAUÁ Case.

Univ. Fed. de Campina Grande, Brazil;
DOI: 10.1109/SBAC-PAD.2004.37 Conference: 16th Symposium on Computer Architecture and High Performance Computing (SBAC-PAD 2004), 27-29 October 2004, Foz do Iguacu, Brazil
Source: DBLP

ABSTRACT In this paper we discuss the difficulties involved in the scheduling of applications on computational grids. We highlight two main sources of difficulties: 1) the size of the grid rules out the possibility of using a centralized scheduler; 2) since resources are managed by different parties, the scheduler must consider several different policies. Thus, we argue that scheduling applications on a grid require the orchestration of several schedulers, with possibly conflicting goals. We discuss how we have addressed this issue in the context of PAUA, a grid for Bag-of-Tasks applications (i.e. parallel applications whose tasks are independent) that we are currently deploying throughout Brazil.

1 Bookmark
 · 
81 Views
  • [Show abstract] [Hide abstract]
    ABSTRACT: A 'Grid' is an infrastructure for resource sharing. It is used for large-scale data processing, many of the applications being scientific ones. The processing of a job on a conventional network takes place on a single computer, whereas the processing of a job on a Grid may be split between several computers. A Grid is, therefore, capable of running applications that are too demanding for a stand-alone computer or a conventional network of computers. The resources making up a Grid need to be managed to provide a good quality of service. Availability is complicated by the fact that a single application may request multiple resource types from multiple locations. Grid resource allocation is regarded as a vital component of a Grid infrastructure. Many methods have been developed for Grid resource allocation. In this paper, we propose a new Grid resource allocation method (Modified Least Cost Method) that assigns tasks to computing nodes in a way that is close to the optimum. The proposed method is an advanced form of the Least Cost Method. In this paper we evaluate Modified Least Cost Method for Grid resource allocation scenario. This paper also proposes a Linear Programming (LP) model for Grid scheduling and includes a comparative performance analysis of our proposed Resource Allocation method with the existing ones.
    Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC), 2010 International Conference on; 01/2010
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Abstract Grid Computing brought the promise of making high-performance computing cheaper and more easily available than traditional supercomputing platforms. Such a promise was very well received by the data mining (DM) community, as DM applications typically process very large datasets and are thus very resource intensive. However, since the Grid is very dynamic and parallel data mining is prone to load unbalancing, obtaining good data mining performance on the Grid is hard. It typically requires for the scheduler to understand the inner works of the application, bringing two related problems. First, good Grid schedulers tend to be very specialized in the application they target. Second, changing the application may require changing the scheduler, what may be specially challenging when there is no clear separation between the application and the scheduler code. We here propose and evaluate a knowledge-based approach that provides abstractions to the DM developer and optimizes at runtime the DM application on the Grid. Keywords: Data Mining Application; Grid; Problem Solving Method; Knowledge-Based System; Ontology.
    Future Generation Comp. Syst. 01/2007; 23:100-106.
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: As the adoption of grid computing in organizations expands, the need for wise utilization of different types of resource also increases. A volatile resource, such as a desktop computer, is a common type of resource found in grids. However, using efficiently other types of resource, such as space-shared resources, represented by parallel supercomputers and clusters of workstations, is extremely important, since they can provide a great amount of computation power. Using space-shared resources in grids is not straightforward since they require jobs a priori to specify some parameters, such as allocation time and amount of processors. Current solutions (e.g. Grid Resource and Allocation Management (GRAM)) are based on the explicit definition of these parameters by the user. On the other hand, good progress has been made in supporting Bag-of-Tasks (BoT) applications on grids. This is a restricted model of parallelism on which tasks do not communicate among themselves, making recovering from failures a simple matter of reexecuting tasks. As such, there is no need to specify a maximum number of resources, or a period of time that resources must be executing the application, such as required by space-shared resources. Besides, this state of affairs makes leverage from space-shared resources hard for BoT applications running on grid. This paper presents an Explicit Allocation Strategy, in which an adaptor automatically fits grid requests to the resource in order to decrease the turn-around time of the application. We compare it with another strategy described in our previous work, called Transparent Allocation Strategy, in which idle nodes of the space-shared resource are donated to the grid. As we shall see, both strategies provide good results. Moreover, they are complementary in the sense that they fulfill different usage roles. The Transparent Allocation Strategy enables a resource owner to raise its utilization by offering cycles that would otherwise go wasted, while protecting the local workload from increased contention. The Explicit Allocation Strategy, conversely, allows a user to benefit from the accesses she has to space-shared resources in the grid, enabling her natively to submit tasks without having to craft (time, processors) requests.
    Future Generation Computer Systems 05/2008; 24(5):331-341. · 1.86 Impact Factor

Full-text (2 Sources)

View
60 Downloads
Available from
May 30, 2014