[WEBINAR] The New Normal: Dealing with the Reality of an Unsecure World

Grid Computing

Definition - What does Grid Computing mean?

Grid computing is a processor architecture that combines computer resources from various domains to reach a main objective. In grid computing, the computers on the network can work on a task together, thus functioning as a supercomputer.

Typically, a grid works on various tasks within a network, but it is also capable of working on specialized applications. It is designed to solve problems that are too big for a supercomputer while maintaining the flexibility to process numerous smaller problems. Computing grids deliver a multiuser infrastructure that accommodates the discontinuous demands of large information processing.

Techopedia explains Grid Computing

A grid is connected by parallel nodes that form a computer cluster, which runs on an operating system, Linux or free software. The cluster can vary in size from a small work station to several networks. The technology is applied to a wide range of applications, such as mathematical, scientific or educational tasks through several computing resources. It is often used in structural analysis, Web services such as ATM banking, back-office infrastructures, and scientific or marketing research.

The idea of grid computing was first established in the early 1990s by Carl Kesselman, Ian Foster and Steve Tuecke. They developed the Globus Toolkit standard, which included grids for data storage management, data processing and intensive computation management.

Grid computing is made up of applications used for computational computer problems that are connected in a parallel networking environment. It connects each PC and combines information to form one application that is computation-intensive.

Grids have a variety of resources based on diverse software and hardware structures, computer languages, and frameworks, either in a network or by using open standards with specific guidelines to achieve a common goal.

Grid operations are generally classified into two categories:

  • Data Grid: A system that handles large distributed data sets used for data management and controlled user sharing. It creates virtual environments that support dispersed and organized research. The Southern California Earthquake Center is an example of a data grid; it uses a middle software system that creates a digital library, a dispersed file system and continuing archive.
  • CPU Scavenging Grids: A cycle-scavenging system that moves projects from one PC to another as needed. A familiar CPU scavenging grid is the search for extraterrestrial intelligence computation, which includes more than three million computers.

Grid computing is standardized by the Global Grid Forum and applied by the Globus Alliance using the Globus Toolkit, the de facto standard for grid middleware that includes various application components.

Grid architecture applies Global Grid Forum-defined protocol that includes the following:

  • Grid security infrastructure
  • Monitoring and discovery service
  • Grid resource allocation and management protocol
  • Global access to secondary storage and GridFTP

Techopedia Deals

Connect with us

Techopedia on Linkedin
Techopedia on Linkedin
"Techopedia" on Twitter

Sign up for Techopedia's Free Newsletter!

Email Newsletter

Join thousands of others with our weekly newsletter

Free Whitepaper: The Path to Hybrid Cloud
Free Whitepaper: The Path to Hybrid Cloud:
The Path to Hybrid Cloud: Intelligent Bursting To Amazon Web Services & Microsoft Azure
Free E-Book: Public Cloud Guide
Free E-Book: Public Cloud Guide:
This white paper is for leaders of Operations, Engineering, or Infrastructure teams who are creating or executing an IT roadmap.
Free Tool: Virtual Health Monitor
Free Tool: Virtual Health Monitor:
Virtual Health Monitor is a free virtualization monitoring and reporting tool for VMware, Hyper-V, RHEV, and XenServer environments.
Free 30 Day Trial – Turbonomic
Free 30 Day Trial – Turbonomic:
Turbonomic delivers an autonomic platform where virtual and cloud environments self-manage in real-time to assure application performance.