Documentos de Académico
Documentos de Profesional
Documentos de Cultura
Artificial Intelligence
and Grids: Workflow
Planning and Beyond
Yolanda Gil, Ewa Deelman, Jim Blythe, Carl Kesselman, and Hongsuda
Tangmunarunkit, USC Information Sciences Institute
geophysics, earthquake engineering, biology, and global climate change (see the “Grid
management, and sharing, grids let geographically This might span several days and necessitate failure
distributed teams form dynamic, multi-institutional handling and reconfiguration to handle the dynam-
A key challenge for grid virtual organizations whose members use shared ics of the grid execution environment.
community and private resources to collaborate on Second, we can significantly multiply the impact
computing is creating solutions to common problems. This gives scien- of scientific research if we broaden the range of
tists tremendous connectivity across traditional orga- applications that it can support beyond science-
large-scale, end-to-end nizations and fosters cross-disciplinary, large-scale related uses. The challenge is to make these complex
research. Grids’ most tangible impact to date could scientific applications accessible to the many poten-
scientific applications be the seamless integration of and access to high- tial users outside the scientific community. In earth-
performance computing resources, large-scale data quake science, for example, integrated earth sciences
that draw from pools sets, and instruments that form the basis of advanced research for complex probabilistic seismic hazard
scientific discovery. However, scientists now pose analysis can have greater impact, especially when it
of specialized scientific new challenges that will require the current grid com- can help mitigate the effects of earthquakes in pop-
puting paradigm to shift significantly. ulated areas. In this case, users might also include
components to derive First, science could progress significantly via the safety officials, insurance agents, and civil engineers,
synthesis of models, theories, and data contributed who must evaluate the risk of earthquakes of certain
elaborate new results. across disciplines and organizations. The challenge magnitude at potential sites. A clear need exists to
is to enable on-demand synthesis of large-scale, end- isolate end users from the complex requirements nec-
Given a user’s high- to-end scientific applications that draw from pools of essary for setting up earthquake simulations and exe-
specialized scientific components to derive elaborate cuting them seamlessly over the Grid.
level specification of new results. Consider, for example, a physics-related We are developing Pegasus, a planning system
application for the Laser Interferometer Gravita- we’ve integrated into the Grid environment that
desired results, the tional-Wave Observatory (LIGO),4 where instru- takes a user’s highly specified desired results,
ments collect data that scientists must analyze to generates valid workflows that take into account
Pegasus system can detect the gravitational waves predicted by Einstein’s available resources, and submits the workflows
theory of relativity (see the “Searching for Gravita- for execution on the Grid. We are also beginning
generate executable tional Waves” sidebar). To do this, scientists run pul- to extend it as a more distributed and knowledge-
sar searches in certain areas of the sky for a certain rich architecture.
grid workflows time period. The observations are processed through
Fourier transforms and frequency range extraction Challenges for robust workflow
using AI planning software. The analysis could involve composing a generation and management
workflow comprising hundreds of jobs and execut- To develop scalable, robust mechanisms that
techniques. ing them on appropriate grid computing resources. address the complex grid applications that the scien-
tific community envisions, we need expres- Clearly, a more flexible and knowledge-rich sophisticated planning and scheduling deci-
sive and extensible ways of describing the Grid infrastructure is needed. Specifically, we sions. Something as central to the Grid as
Grid at all levels. We also need flexible mech- must address the following issues. resource descriptions are still based on rigid
anisms to explore trade-offs in the Grid’s com- schemas. Although higher-level middleware
plex decision space that incorporate heuristics Knowledge capture is under development,2,5 grids will have a
and constraints into that process. In contrast, High-level services such as workflow gen- performance ceiling determined by limited
Grids today use syntax or schema-based eration and management systems are starved expressivity and the amount of information
resource matchmakers, algorithmic sched- for information and lack expressive descrip- and knowledge available for making intelli-
ulers, and execution monitors for scripted job tions of grid entities and their relationships, gent decisions.
sequences, which attempt to make decisions capabilities, and trade-offs. Current grid mid-
with limited information about a large, dleware simply doesn’t provide the expres- Usability
dynamic, and complex decision space. sivity and flexibility necessary for making Exploiting distributed heterogeneous
Resource
Resource knowledge
indexes base
Application
knowledge
Workflow Workflow base
refinement Workflow
history
Workflow
history Simulation
Policy management history Replica
locators codes
Smart workflow
Pool Community distributed resources Other
Resource (for example, computers, storage, grid
matching Workflow Workflow manager network, simulation codes, data) services
repair Policy
knowledge Policy
base information
services
Other
knowledge
Intelligent base
Pervasive
reasoners knowledge
sources
a representation of the current state of its between hosts performing related tasks. Future grid workflow
environment, a declarative representation of Additionally, planning techniques can pro- management
a goal state, and a library of operators that vide high-quality solutions, partly because We envision many distributed heteroge-
the planner can use to change the state. Each they can search several solutions and return neous knowledge sources and reasoners, as
operator has a description of the states in the best ones found, and because they use Figure 1 shows. The current grid environ-
which it can legally be used, called precon- heuristics that will likely guide the search ment contains middleware that can find com-
ditions, and a concise description of the to good solutions. ponents that can generate desired results, find
changes to the state that will take place, Pegasus takes a request from the user and the input data they require, find replicas of
called effects. The planning system searches builds a goal and relevant initial state for the component files in specific locations, match
for a valid, partially ordered set of operators AI planner, using grid services to locate rel- component requirements with available
that will transform the current state into one evant existing files. Once the plan is com- resources, and so on. This environment
that satisfies the goal. Each operator’s para- plete, Pegasus transforms it into a directed should be extended with expressive declara-
meters include the host where the component acyclic graph to pass to DAGMan11 for exe- tive representations that capture currently
is to run, while the preconditions include cution on the Grid. implicit knowledge, and should be available
constraints on feasible hosts and data depen- We are using Pegasus to generate exe- to various reasoners distributed throughout
dencies on required input files. The plan cutable grid workflows in several domains,7 the Grid.
returned corresponds to an executable work- including genomics, neural tomography, and In our view, workflow managers will coor-
flow, which includes the assignment of com- particle physics (see the LIGO application dinate the generation and execution of work-
ponents to specific resources that can be exe- description in the “Searching for Gravita- flow pools. The workflow managers’ main
cuted to provide the requested data product. tional Waves” sidebar). responsibilities are to
The declarative representation of actions As we attempt to address more aspects of
and search control in domain-independent the grid environment’s workflow management • Oversee their assigned workflows’ devel-
planners is convenient for representing con- problem, including failure recovery, respect- opment and execution
straints such as computation and storage ing institutional and user policies and prefer- • Coordinate among workflows that might
resource access and usage policies. Plan- ences, and optimizing various global measures, have common subtasks or goals
ners can also incorporate heuristics, such we find that, as mentioned, a more distributed • Apply fairness rules to ensure that work-
as preferring a high-bandwidth connection and knowledge-rich approach is required. flows execute in a timely manner
To Submit a Manuscript: To submit a manuscript for peer-reviewed consideration, please access the IEEE Computer Society
Web-based system, Manuscript Central, at http://cs-ieee.manuscriptcentral.com/index.html.
M
anufacturing organizations face conditions of unprecedented ing intelligent control systems for the manufacturing supply chain. In
disruption and change, and systems that control the many particular, the issue will aim to position this work in terms of its poten-
operations along the manufacturing supply chain must be tial longer-term impact on industry and on the issues required to see
able to adapt to these conditions. Recently,a major thrust in address- more widespread deployment. In addition, the issue will explain some
ing these requirements has been the application of tools from distrib- fundamental research concepts in this area.
uted AI. The deployment of tools such as neural networks, fuzzy logic,
and evolutionary programming has provided new routes for tackling For this special issue, we invite original, high-quality submissions that
complex issues in scheduling and control of manufacturing processes. address all aspects of intelligent control as it is applied to the manufac-
In addition, manufacturing control and management systems based turing supply chain. Submissions must address the issues of how the
on the multiagent system paradigm have received significant attention, developments described will impact the manufacturing supply chain
because they promise to provide a high flexibility and easy reconfig- and what barriers to their adoption exist. Papers addressing perfor-
urability in the face of changes. A closely related development is the mance evaluation of intelligent control systems versus more conven-
holonic manufacturing systems methodology, which couples intelligent tional systems will be extremely welcome.
software elements such as agents with physical entities such as equip-
ment, orders, and products to effectively provide a “plug and play”
factory. Many of these developments are now at the point where in-
dustrial deployment is a serious possibility and major systems vendors
are considering integrating intelligent control capabilities into their Guest Editors
product offerings. Duncan McFarlane, University of Cambridge
Vladimir Marik, Czech Technical University
This special issue will feature articles that address the issue of develop-
Paul Valckenaers, Katholieke Universiteit Leuven
Ewa Deelman is a research team leader at the Center for Grid Technologies
at the University of Southern California’s Information Sciences Institute and
nected information and computing infra- an assistant research professor at the USC Computer Science Department.
structure that will harness the power and Her research interests include the design and exploration of collaborative sci-
diversity of massive amounts of online sci- entific environments based on grid technologies, with particular emphasis on
entific resources. Our work contributes to this workflow management as well as the management of large amounts of data
and metadata. At ISI she is part of the Globus project, which designs and imple-
vision by addressing two central questions: ments Grid middleware. She received her PhD from Rensselaer Polytechnic
Institute in Computer Science in 1997 in the area of parallel discrete event
• What mechanisms can map high-level simulation. Contact her at USC/ISI, 4676 Admiralty Way, Marina del Rey, CA 90292; deelman@isi.edu.
requirements from users into distributed
Jim Blythe is a computer scientist in USC’s Information Sciences Institute.
executable commands that pull numerous His research interests include planning, grid workflows, and knowledge acqui-
distributed heterogeneous services and sition for problem solving and planning applications. He received his PhD in
resources with appropriate capabilities to computer science from Carnegie Mellon University. Contact him at the
meet those requirements? ISI/USC, 4676 Admiralty Way, Marina del Rey, CA 90292-6696;
• What mechanisms can manage and coor- blythe@isi.edu.
dinate the available resources to enable
efficient global use and access given the
scale and complexity of the applications Carl Kesselman is a fellow in USC’s Information Sciences Institute. He is
possible with this highly distributed het- the director of the Center for Grid Technologies at ISI, and a research asso-
erogeneous infrastructure? ciate professor of computer science at USC. He received his PhD in com-
puter science from the University of California, Los Angeles. Together with
Ian Foster, he cofounded the Globus Project, a leading Grid research project
The result will be a new generation of scien- that has developed the Globus Toolkit, the defacto standard for Grid com-
tific environments that can integrate diverse puting. Contact him at USC/ISI, 4676 Admiralty Way, Marina del Rey, CA
scientific results and whose sum will be 90292; carl@isi.edu.
orders of magnitude more powerful than their
individual ingredients. The implications will Hongsuda Tangmunarunkit is a computer scientist in the Center for Grid
go beyond science and into the realm of the Technologies at USC’s Information Sciences Institute. Her research interests
Web at large. include grid computing, networking, and distributed systems. She received
her MS in computer science from USC. She is a member of the ACM. Con-
tact her at USC/ISI, 4676 Admiralty Way, Suite 1001, Marina del Rey, CA
Acknowledgments 90292; hongsuda@isi.edu; www.isi.edu/~hongsuda.
We thank Gaurang Mehta, Gurmeet Singh, and
Karan Vahi for developing the Pegasus system. We
also thank Adam Arbree, Kent Blackburn, Richard
Cavanaugh, Albert Lazzarini, and Scott Koranda.
The visualization of LIGO data was created by Mar- 4. A. Abramovici et al., “LIGO: The Laser Inter- 10. J. Blythe et al., “Transparent Grid Comput-
cus Thiebaux using a picture from the Two Micron ferometer Gravitational-Wave Observatory ing: A Knowledge-Based Approach,” Proc.
All Sky Survey NASA collection. This research was (in Large Scale Measurements),” Science, vol. 15th Innovative Applications of Artificial
supported partly by the National Science Founda- 256, 1992, pp. 325–333. Intelligence Conf. (IAAI), AAAI Press, 2003,
tion under grants ITR-0086044 (GriPhyN) and pp. 57–64.
EAR-0122464 (SCEC/ITR), and partly by an inter- 5. T.H. Jordan et al., “The SCEC Community
nal grant from the Information Sciences Institute. Modeling Environment—An Information 11. J. Frey et al., “Condor-G: A Computation Man-
Infrastructure for System-Level Earthquake agement Agent for Multi-institutional Grids,”
Research,” Seismological Research Letters, Cluster Computing, vol. 5, 2002, pp. 237–246.
vol. 74, no. 3, May/June 2003, pp. 44–46.
References 12. I. Foster, C Kesselman, and S. Tuecke, “The
6. E. Deelman et al., “Workflow Management Anatomy of the Grid: Enabling Scalable Vir-
1. J. Annis et al., Applying Chimera Virtual Data in GriPhyN,” Grid Resource Management, J. tual Organizations,” Int’l J. Supercomputer
Concepts to Cluster Finding in the Sloan Sky Nabrzyski, J. Schopf, and J. Weglarz, eds., Applications, vol. 15, no. 2, 2001; www.globus.
Survey, tech. report GriPhyN-2002-05, Grid Kluwer, 2003. org/research/papers/anatomy.pdf.
Physics Network, 2002; www.griphyn.org/
chimera/papers/SDSS-SC2002.crc. 7. Ewa Deelman et al., “Pegasus: Mapping Sci- 13. O. Lassila and R.R. Swick, Resource Descrip-
submitted.pdf. entific Workflows onto the Grid,” to be pub- tion Framework (RDF) Model and Syntax Spec-
lished in Proc. 2nd European Across Grids ification, W3C Recommendation, World Wide
2. P. Avery and I. Foster, The GriPhyN Project: Conf., 2004. Web Consortium, 22 Feb. 1999; www.w3.
Towards Petascale Virtual Data Grids, tech. org/TR/1999/REC-rdf-syntax-19990222.
report GriPhyN-2001-15, Grid Physics Net- 8. E. Deelman et al., “Mapping Abstract Com-
work, 2001. plex Workflows onto Grid Environments,” J. 14. E. Christensen et al., Web Services Description
Grid Computing, vol. 1, no. 1, 2003, pp. 25–39. Language (WSDL) 1.1, World Wide Web Con-
3. E. Deelman et al., “GriPhyN and LIGO, sortium, 15 Mar. 2001; www.w3.org/TR/wsdl.
Building a Virtual Data Grid for Gravitational 9. J. Blythe et al., “The Role of Planning in Grid
Wave Scientists,” Proc. 11th Int’l Symp. High- Computing,” Proc. 13th Int’l Conf. Automated For more information on this or any other com-
Performance Distributed Computing, IEEE Planning and Scheduling (ICAPS 2003), puter topic, please visit our Digital Library at
CS Press, 2002. AAAI Press, 2003, pp. 153–163. www.computer.org/publications/dlib.