Академический Документы
Профессиональный Документы
Культура Документы
Jon Wakelin
University of Bristol
Pre-Amble
GPFS Basics
What it is & what it does
GPFS Concepts
More in-depth technical concepts
GPFS Topologies
StoRM
Recap & References
GPFS Basics
IBMs General Parallel File System
Scaleable high-performance parallel file system
Numerous HA features
Life-cycle Management Tools
Provides POSIX and extended interfaces to data
Multi-Cluster Configuration
Join GPFS clusters together
Encrypted data and authentication or just authentication
openssl and keys
Different security contexts (root squash la NFS)
GPFS Basics
Information Life-cycle Management
Tiered storage
Create groups of disks within a file system,
based on reliability, performance, location, etc
Policy driven automation
Automatically move, delete or replicate files - based on filename, username, or fileset.
e.g. Keep newest files on fastest hardware, migrate them to older hardware over time
e.g. Direct files to appropriate resource upon creation.
SAN-Attached
All nodes are physically attached to all NSDs
High performance but expensive!
GPFS Topologies
HA Features in GPFS
Primary and secondary Configuration Servers
Primary and secondary NSD Servers for each Disk
Replicate Metadata
Replicate data
Failure Groups
Specify which machines have a single point of failure
GPFS will use this info to make sure that replicated data is not striped across failure groups
GPFS Quorum
Quorum
A Majority of the nodes must be present before access to shared disks is allowed
Prevent subgroups making conflicting decisions
In event of failure disks in minority suspend and those in the majority continue
Quorum Nodes
These nodes are counted to determine if the system is quorate
If the system is no longer quorate
GPFS unmounts the filesystem
waits until quorum is established
and then recovers the FS.
1 process 90MB/s
2 processes 51 MB/s
4 processes 18MB/s
1 process 90MB/s
2 processes 58 MB/s
4 processes 28 MB/s
5 processes 23 MB/s
GPFS Performance
In a hybrid environment (SAN-attached and NSD Server nodes)
Read/Writes from SAN-attached nodes place little load on the NSD servers
Read/Writes from other nodes place a high load on the NSD servers
SAN-attached
[root@bf39 gpfs]# time dd if=/dev/zero of=file_zero count=2048 bs=1024k
real 0m31.773s
Testing
A number of pump-priming projects have been identified
Majority of users will develop, or port code, directly on the HPC system
Only make changes at the Application level
GridPP
System level changes
Pool accounts, World-addressable Slaves, NAT, Run services and daemons
Server Room
48 water cooled APC racks 18 will be occupied by HPC, Physics Servers may be co-
located
3 x270kW chillers (space for 3 more)
GPFS BabyBlue
GPFS MiniBlue
IBM DS4500
Configure hot spares
StoRM
StoRM is a storage resource manager for disk based storage systems.
Implements the SRM interface version 2.2
StoRM is designed to support guaranteed space reservation and direct access (using native
POSIX I/O call)
StoRM takes advantage of high performance parallel file systems
GPFS, XFS and Lustre???
Also standard POSIX file systems are supported
Direct access to files from Worker Nodes
Compare with Castor, D-Cache and DPM
StoRM architecture
Front end (FE):
Exposes the web service interface
Manages user authentication
Sends the request to the BE
HPC
University wide facility not just for PP
GridPP requirements rather different from general/traditional HPC users
Build an analogue of the HPC system for GridPP
StoRM
Better performance because StoRM builds on
Also, more appropriate data transfer model POSIX and file protocol
References
GPFS
http://www-03.ibm.com/systems/clusters/software/gpfs.pdf
http://publib.boulder.ibm.com/infocenter/clresctr/vxrx/topic/com.ibm.cluster.gpfs.doc/gpfs_fa
qs/gpfsclustersfaq.pdf
http://www-03.ibm.com/systems/clusters/software/whitepapers/gpfs_intro.pdf
Storm
http://hst.home.cern.ch/hst/publications/storm_chep06.pdf
http://agenda.cnaf.infn.it/getFile.py/access?contribId=10&resId=1&materialId=slides&confId=0