This year marks the 30th anniversary of the birth of the Internet, which
originally linked universities and laboratories around the United States as
an outgrowth of a Defense Department project begun in 1969.
How the baby has grown! The Internet began to expand from its inception,
and with the introduction of the World Wide Web in 1989, it is becoming
ubiquitous. It will be as much a part of people’s lives as the radio,
television, and telephone — all of which will no doubt reside on the
Internet in the not-too-distant future.
The Internet’s explosive growth has been fueled by a single factor: open
standards — an agreement upon a common set of “protocols” enabling
computers to communicate with one another. Standards also brought Internet
mail into existence, which has had an incalculable impact upon the way we
communicate with each other. And yet another set of standards created the
Web, which has allowed everyone with a computer to trade information.
It did not take long for business to see the potential of the Internet.
Today you’d be hard-pressed to find a company that does not use the Internet
for emailing sales information to customers, or displaying Website
“showrooms,” or integrating sales and supply-chain activities in order to
maintain optimal inventory levels.
As exciting as the Internet’s past 30 years have been, they’re only a shadow
of what’s to come. We’re already catching glimpses of the future in
projects sponsored by universities and pharmaceutical research labs. By
downloading a special screen saver, we can allow our personal computers to
hook into a worldwide network when we’re not using the machines. Together,
this network processes information as part of the search to find cures for
cancer, HIV, and, most recently, smallpox — processing that would take an
individual computer weeks or months, but can be done in days through this
In a somewhat different form, grid computing — a Web-based operation that
will allow companies to share computing resources on demand — is where
we’re headed. Simply put, grid computing uses more of a server’s computing
power. Today’s computers, like human brains, typically operate at only a
small percentage of their capacity; they often sit idle as a processor waits
for data. On the grid, the idle time of hundreds — even thousands — of
servers can be harnessed by any customer needing a massive infusion of
Just as the electric grid provides power to the consumer, the computing grid
distributes all sorts of computing resources to solve problems. In the
process, a grid can give rise to “virtual organizations” — ever-changing
groups of individuals and institutions exploiting the resources of the grid
for a variety of purposes, much the way that individuals in the same
household exploit electricity for their own needs.
Grid computing is the logical, and desirable, outcome of several factors.
First, the idle time inherent in computing means that by working on a grid,
computers can share operations according to their abilities and unused
capacities; thus, information will be processed with maximum efficiency.
Second, the advent of broadband enables networked computers to share data
constantly at high speeds — critical to the functioning of any
Of greatest importance, though, is the collaborative culture that the
Internet gave us in the first place. Without common protocols, the Web
would be a Tower of Babel. Businesses, and the companies that supply
technology to business, are understanding that the virtues of open standards
— flexibility, scalability, and adaptability — are keys to the success of
information technology systems and to the ability of those systems to
contribute to an attractive bottom line. Businesses are thus developing IT
systems that embrace open standards, and are constructing those systems
around “middleware”: software that enables different technologies to
interact with one another smoothly and transparently.
Capacity, speed, and open systems are setting the stage for the leap to grid
computing — for the Internet (or an intranet) itself to become a computing
platform. The resources we desire will be at our fingertips, but they will
reside on the Internet rather than in single computers or on local servers.
We will engage these resources through secure channels, as we do now when we
transact business on the Internet. But we will never again wish that we had
more storage capacity, more data-manipulation capabilities, or another suite
of applications. The Internet’s “virtual computer” will be distributed
worldwide rather than existing in one place. It will provide us with
everything we need, whenever we need it, and will enable us to coordinate
our work with anyone, anywhere, without technology conflicts. Like power
grids, computer grids will become an essential element of our
infrastructure, operating invisibly except in the rare cases when a problem
Like Linux and other new technologies, grid computing is emerging first in
the scientific and technical communities, enabling scientists and engineers
to collaborate in applications across institutions and around the world in
disciplines such as high-energy physics, life sciences, and engineering
design. For example, the Mayo Clinic is developing a system for linking its
own medical database with vast external public and private data sources in
order to develop more-effective patient treatments. The University of
Pennsylvania is building a powerful computing grid that will make remote
breast cancer diagnosis and screening a reality, putting the benefits of
sophisticated, high-tech healthcare delivery at the disposal of entire
populations. And Indiana and Purdue universities have linked their systems
and developed an application that helps assess the impact of catastrophic
events on populations.
The next step is for grid computing to break into the mainline business
world. Butterfly.net uses a grid for its massive, multi-player online
games. And grids are beginning to see use by financial institutions for
applications such as risk-management analysis and portfolio optimization, by
manufacturing companies for engineering simulation, and in the energy
industry for seismic analysis.
Of course, Utopia has yet to arrive, but these early trials hold the promise
of the future. New standards, beyond interoperability, must emerge among
different computers and networks. And if we are to make computing grids an
essential, yet invisible, part of our society, we must focus on simplicity
of design and create a self-healing, “autonomic” infrastructure. This is an
architectural style popularized by the “Blue Gene” supercomputer, which is
being developed to give scientists and doctors greater insight into
combating diseases. Blue Gene has eight million autonomous computing
elements that cooperate among each other and self-manage the computations.
The promise of grid computing — of Web-based services providing universal
access to information and computing in a collaborative environment — is as
real as it is seductive. In fact, it’s only fitting that the Internet, born
out of a vision of collaboration, should now come full circle to fulfill
About the Author
Dr. Willy Chiu is the Vice President of the High- Volume Web sites
organization in the IBM Software Group. He and his team of experts have
created a world class center of competence on the scalability of high-volume
Web sites; the center serves as a focal point for customers to address the
challenges of excessive growth of Web site traffic and increasing complexity
of e-business infrastructures.
Dr. Chiu’s work helps enterprises benefit from networks that act like the
electrical grid or software that’s able to fix itself. Such innovations are
in line with IBM’s e-business on demand initiative.