Affects
- CompletedMarch 06, 2023 at 5:06 PMCompletedMarch 06, 2023 at 5:06 PM
Maintenance has completed successfully at 12:00PM.
- In progressMarch 06, 2023 at 4:30 PMIn progressMarch 06, 2023 at 4:30 PM
Maintenance is still in progress as of 11:30AM as one of our vendors needs to complete some hardware work.
Access to the Slurm scheduler and all compute are still paused, but access to storage and other services has been restored.
We appreciate your patience.
- CompletedMarch 06, 2023 at 4:00 PMCompletedMarch 06, 2023 at 4:00 PM
Maintenance has completed successfully
- In progressMarch 06, 2023 at 12:00 PMIn progressMarch 06, 2023 at 12:00 PM
Maintenance is now in progress
- PlannedMarch 06, 2023 at 12:00 PMPlannedMarch 06, 2023 at 12:00 PM
NOTICES
The annual MGHPCC power downtime will take place June 5th-8th, 2023
Calendar Event: https://www.rc.fas.harvard.edu/events/mghpcc-power-shutdown-2023/
Blog Post: https://www.rc.fas.harvard.edu/blog/2023-downtime/GENERAL MAINTENANCE
NOTE: All jobs will be paused during maintenance to reduce heat load and allow data center cooling maintenance to take place.
Login node updates and reboots, VDI reboots
Audience: VDI/OpenOnDemand users
Impact: VDI will be unavailable during this and the above Slurm upgradeRCSMB (samba) Boston network changes
Audience: RCSMB shares mounted out of Boston
Impact: Could cause brief share disconnects during updatesUPDATE: Nexus control plane supervisor switchover - ETA 5 minutes, short network disconnect while restarting
Login node updates/reboot and VDI node reboots
Audience: Anyone logged into a a login node or VDI/OOD node
Impact: Login and VDI/OOD nodes will be unavailable while updating and rebootingScratch cleanup ( https://docs.rc.fas.harvard.edu/kb/policy-scratch/ )
Audience: Cluster users
Impact: Files older than 90 days will be removed.
Reminder: Scratch 90-day file retention purging runs occur regularly not just during maintenance periods.
SECURITY UPDATES
HUIT and the CIO Council have set a goal of reducing risk across all schools within the University. All schools are looking to reduce their outstanding vulnerability count 75% by June 2023. These numbers are based on HUIT security scans of our infrastructure.
We at FAS Research Computing are responsible for thousands of physical and virtual machines. To make progress in reducing our total open vulnerabilty count, we’re going to update internal and user facing systems as part of scheduled monthly maintenance windows and on a rolling basis outside of these windows. This will generally mean running OS and security updates as needed and rebooting these nodes when required.This month, these hosts will get updates as part of our scheduled maintenance:
boslogin01 - boslogin04
holylogin01 - holylogin04
holydtn01 - holydtn04
xdmod4.rc.fas.harvard.edu
rchelp.rc.fas.harvard.edu (our ticket system)
rcsmtp.rc.fas.harvard.edu (our mail system)Thanks!
FAS Research Computing
Department and Service Catalog: https://www.rc.fas.harvard.edu/
Documentation: https://docs.rc.fas.harvard.edu/
Status Page: https://status.rc.fas.harvard.edu/