Академический Документы
Профессиональный Документы
Культура Документы
INTRODUCTION
The initial applications were recently built following different distributed computing approaches:
(1)A Service Oriented Architecture (SOA) based training system
(2) A modular real-time data analyzer
(3) A cluster-based simulator.
But Cloud technologies are currently designed mainly for developing new applications. Early
Cloud providers were focused on developers and technology startups when they designed their
offerings. Software architects looking to build new applications can design the components,
processes and workflow for their solution according to the new Cloud related concepts.
However, building new applications that are being architected from scratch for the Cloud is only
slowly gaining traction, and there are only few enterprise applications that currently take real
advantage of the Clouds elasticity. Having an application distributed across multiple clouds to a
large extent reduces the risk of data security and storage, as well as power and equipment
breakdown. This is one of the reasons that led to the bringing together of several clouds (owned
by different providers) to form what is known as Sky Computing. Infrastructure as- a-service
(IaaS) cloud computing is revolutionizing how we approach computing. Infrastructure-as-a-
service (IaaS) cloud computing represents a fundamental change from the grid computing
Reconciling those choices between multiple user groups proved to be complex, time consuming,
and expensive. Compute resource consumers can eliminate the expense inherent in acquiring,
managing, and operating IT infrastructure and instead lease resources on a pay-as-you-go basis.
IT infrastructure providers can exploit economies of scale to mitigate the cost of buying and
operating resources and avoid the complexity required to manage multiple customer-specific
environments and applications. So, this complexity helped to arose an emerging computing
pattern known as Sky Computing. The main advantage of the cloud computing is that this
technology reduces the cost effectiveness for the implementation of the Hardware, software and
License for all Users can further benefit from low cost and high resource utilization by using sky
computing.
To achieve this, there must be a structure capable of receiving instructions, process and return
results from all different underlying cloud systems. The upper layer, Sky Computing, integrates
the last level of Infrastructure as a Service and the next layer of Software as a Service. It allows
scheduling and distributing resources to inputted tasks/requests made. This is a critical layer, as it
must be as comprehensive as possible in features and capabilities.
End-user clusters: In the absence of ViNe services from providers, users can enable
ViNe as an additional VM that they start and configure to connect different cloud providers. This
user deployed VR would handle the traffic crossing the cluster nodes LAN boundaries. ViNes
cost in this case is an additional VM per user.
Isolated VMs: A VR cant be used as a gateway by machines that dont belong to the
same LAN segment. In this case, every isolated VM (or a physical machine, such as the users
client machine) must become a VR. ViNes cost is the additional network processing that
compute nodes perform, which can take compute cycles from applications.
End users provide context information via a simple generic schema and method thats the
same for every appliance used with this provider. Adopting this simple schema lets every
provider deliver basic context information to every appliance.
Appliance providers, who provide methods that let appliance supply information to and
receive it from a context broker and integrate information conveyed by templates describing
Fig
4: A virtual
cluster
We deployed the two virtual clusters in two settings: on the UF cloud only (one-site experiment)
and on all three sites using the same number of processors (three-site experiment). For three-site
experiments, we balanced the number of hosts in each site executing Blast that is, one host in
each site, two hosts in each site, and so on, up to five hosts in each site. (Choosing random
numbers of nodes from different sites would, in effect, weigh the three-site experiments
performance toward comparing the UF site and the site with the most processors). The SLAs
expressed as instances from each metacloud provider are different (PU instances outperform UC
instances which outperform UF instances), which makes it difficult to compare providers. To
establish a comparison base between the SLAs each provider offers, we used the performance of
the sequential execution on a UF processor of the Blast job described earlier to define a
normalized performance benchmark 1 UC processor is equivalent to 1.184 UF processors,
whereas 1 PU processor is equivalent to 1.24 UF processors. For example, an experiment with 10
UF processors, 10 UC processors, and 10 PU processors should provide the performance of a
cluster with 34.24 UF processors. We used these factors to normalize the number of processors
Figure 4 shows the speedup Blast execution on various numbers of testbed processors in
different deployment settings versus the execution on one processor at UF. A sequential
execution on one UF processor resource that took 43 hours and 6 minutes was reduced to 1 hour
and 42 minutes using Hadoop on 15 instances (30 processors) of the UF cloud, a 25.4-fold
CHAPTER-3
GRID AND CLOUD COMPUTING
Agent Agent
Cooperation
Agent Cooperation
Distribution Distribution Cooperation
Distribution
Internet
Internet
Agent
Subscription Distribution
Job Request
Resource Large-scale
Management Application
In the term distributed computing, the word distributed means spread out across space.
Thus, distributed computing is an activity performed on a spatially distributed system. These
Characteristics:
Resource Sharing
Openness
Concurrency
Scalability
Fault Tolerance
Transparency
Architecture:
Client-server
3-tier architecture
N-tier architecture
Peer-to-peer
Space based
Mobile agents can be wandering around in a network using free resources for their own
computations
Local intranet
Benefits
Grid Applications
Resource sharing
Grid Topologies
Intragrid
Extragrid
Intergrid
COMPUTATIONAL GRID
The Grid: Blueprint for a New Computing Infrastructure, Kesselman & Foster
DATA GRID
A data grid is a grid computing system that deals with data the controlled sharing and
management of large amounts of distributed data.
Data Grid is the storage component of a grid environment. Scientific and engineering
applications require access to large amounts of data, and often this data is widely
distributed. A data grid provides seamless access to the local or remote data required to
complete compute intensive calculations.
Example:
Distributed Supercomputing
High-Throughput Computing
On-Demand Computing
Data-Intensive Computing
Collaborative Computing
Logistical Networking
Distributed Supercomputing
High-Throughput Computing
On-Demand Computing
Uses grid capabilities to meet short-term requirements for resources that are not locally
accessible.
Models real-time computing demands.
Collaborative Computing
Data-Intensive Computing
Logistical Networking
Contrasts with traditional networking, which does not explicitly model storage resources
in the network.
IP hourglass model
Collective (app): Coherency control, replica selection, task management, virtual data
catalog, virtual data code catalog
Bricks scheduling
GangSim- Ganglia VO
All components in GridSim communicate with each other through message passing
operations defined by SimJava.
Resource capability can be defined (in the form of MIPS (Million Instructions per
Second) benchmark.
Weekends and holidays can be mapped depending on resources local time to model non-
Grid (local) workload.
Application tasks can be heterogeneous and they can be CPU or I/O intensive.
There is no limit on the number of application jobs that can be submitted to a resource.
Multiple user entities can submit tasks for execution simultaneously in the same resource,
which may be time-shared or space-shared. This feature helps in building schedulers that
can use different market-driven economic models for selecting services competitively.
CHAPTER-4
A method of design, deployment, and management of both applications and the software
infrastructure where:
All software is organized into business services that are network accessible and
executable.
Software: 1 - 4 years
Hardware: 3 - 5 years;
Communications: 1 - 3 years;
Data: forever.
2) Data Services
3) Security Services
4) Computing Services
5) Communication Services
6) Application Services
Data Concepts
Text associated with a unique data element within a data dictionary that describes
the data element, give it a specific meaning and differentiates it from other data
elements. Definition is precise, concise, non-circular, and unambiguous.
Security Services
Computing Services
Communication Services
The Lightweight Directory Access Protocol, or LDAP is protocol for querying and
modifying directory services.
Extract, Transform, and Load, ETL, is a process of moving data from a legacy system
and loading it into a SOA application.
CHAPTER-5
REFERENCES