We are seeing a reduction in the GPFS filesystem problems over the past weekend, and are continuing to actively work with the vendor. We don’t have a complete solution yet, but have observed greater stability for compute nodes in the GPFS filesystem. Thank you for your patience – we will continue to keep you updated as much as possible as the situation changes.
Posts
Storage (GPFS) Issue Update
While the problem wasn’t very widespread and we have improved the reliability, we have not yet arrived at a full solution and are still actively working on the problem. We now believe the problem is due to the recent addition of many compute nodes, ultimately bringing us into the next tier of system-level tuning needed for the filesystem. Thank you for your patience – we will continue to provide updates as they become available.
Storage (GPFS) Issue
We are experiencing intermittent problems with the GPFS storage system that hosts scratch and project directories (~/scratch, and ~/data). At the moment, we are exploring this failure with the vendor if this may be related to the recent cluster nodes that have been brought online.
This issue has potential impact on running jobs. We are actively working on the problem, apologize for the inconvenience, and will update as soon as possible.
Storage (GPFS) and datacenter problems resolved
All nodes and GPFS filesystem issues affected from the power failure should be resolved as of late Friday evening (June 16) . If you are still experiencing problems, please let us know at pace-support@oit.gatech.edu.
PACE is experiencing storage (GPFS) problems
We are experiencing intermittent problems with the GPFS storage system that hosts most of the project directories.
We are working with the vendor to investigate the ongoing issues. At this moment we don’t know whether they are related to yesterday’s power/cooling failures or not, but we will update the PACE community as we find out more.
This issue has potential impact on running jobs and we are sorry for this inconvenience.
PACE datacenter experienced a power/cooling failure
Impacted Queues:
Large Scale Problem
Update (6/7/2017, 1:20pm): The network issues are now addressed and systems are back in normal operation.Please check your jobs and resubmit failed jobs as needed. If you continue to experience any problems, or need our assistance for anything else, please contact us at pace-support@oit.gatech.edu. We are sorry for this inconvenience and thank you once again for your patience.
Original message: We are experiencing a large scale network problem impacting multiple storage servers and software repository with a potential impact on running jobs. We are currently actively working to get it resolved and will provide updates as much as possible. We appreciate your patience and understanding, and are committed to resolving the issue as soon as we possibly can.
Infiniband switch failure causing partial network and storage unavailability
The switch is now back online and it’s safe to submit new jobs.
If you are using one or more of the queues (listed below), please check your jobs and re-submit them if necessary. One indication of this issue is “Stale file handle” error messages that may appear in the job output or logs.
Impacted Queues:
=============
athena-intel
atlantis
atlas-6-sunge
atlas-intel
joe-6-intel
test85
apurimacforce-6
b5force-6
bioforce-6
ceeforce
chemprot
cnsforce-6
critcelforce-6
cygnusforce-6
dimerforce-6
eceforce-6
faceoffforce-6
force-6
hygeneforce-6
isblforce-6
iw-shared-6
mathforce-6
mayorlab_force-6
medprint-6
nvidia-gpu
optimusforce-6
prometforce-6
rombergforce
sonarforce-6
spartacusfrc-6
try-6
testflight
novazohar
PACE clusters ready for research
Our May 2017 maintenance period is now complete, far ahead of schedule. We have brought compute nodes online and released previously submitted jobs. Login nodes are accessible and data available. As usual, there are some straggling nodes we will address over the coming days.
Our next maintenance period is scheduled for Thursday, August 10 through Saturday, August 12, 2017.
New operating system kernel
- All compute, interactive, and head nodes have received the updated kernel. No user action needed.
DDN firmware updates
- This update brought low level firmware on drives up to date per recommendation from DDN. No user action needed.
Networking
- DNS/DHCP and firewall updates per vendor recommendation applied by OIT Network Engineering.
- IP address reassignments for some clusters completed. No user action needed.
Electrical
- Power distribution repairs completed by OIT Operations. No user action needed.
PACE quarterly maintenance – May 11, 2017
PACE clusters and systems will be taken offline at 6am this Thursday (May 11) through the the end of Saturday (May 13). Jobs with long walltimes will be held by the scheduler to prevent them from getting killed when we power off the nodes. These jobs will be released as soon as the maintenance activities are complete.
Planned improvements are mostly transparent to users, requiring no user action before or after the maintenance.
Systems
- We will deploy a recompiled kernel that’s identical to the current version except for a patch that addresses the dirty cow vulnerability. Currently, we have mitigation in place that prevents the use of debuggers and profilers (e.g. gdb, strace, Allinea DDT, etc). After the deployment of the patched kernel, these functions will once again be available for all nodes. Please let us know if you continue to have problems debugging or profiling your codes after the maintenance day.
Storage
- Firmware updates on all of the DDN GPFS storage (scratch and most of the project storage)
Network
- Upgrades to DNS servers, as recommended and performed by OIT Network Engineering
- Software upgrades to the PACE firewall appliance to address a known bug
- New subnets and re-assignment of IP addresses for some of the clusters
Power
- PDU fixes that are impacting 3 nodes in c29 rack
The date for the next maintenance day is not certain yet, but we will announce it as soon as we have it.