timestamp1724721093939Storage quota units change: TB to TiBby Kilian Cavalotti, Technical Lead & Architect, HPCFollowing in Oak footsteps, we’re excited to announce that Sherlock is adopting a new unit of measure for file system quotas. Starting today, we're transitioning from Terabytes (TB) to Tebibytes (TiB) for all storage allocations on
timestamp1707349764699Sherlock goes full flashby Stéphane Thiell & Kilian Cavalotti, Research Computing TeamDataHardwareImprovementWhat could be more frustrating than anxiously waiting for your computing job to finish? Slow I/O that makes it take even longer is certainly high on the list. But not anymore! Fir, Sherlock’s scratch file system, has just undergone a major
timestamp1683930644259Final hours announced for the June 2023 SRCF downtimeby Kilian Cavalotti, Technical Lead & Architect, HPCMaintenanceAnnounceAs previously announced, the Stanford Research Computing Facility (SRCF), where Sherlock is hosted, will be powered off during the last week of June, in order to safely bring up power to the new SRCF2 datacenter. Sherlock will not be
timestamp1679706261451A new tool to help optimize job resource requirementsby Kilian Cavalotti, Technical Lead & Architect, HPCIt’s not always easy to determine the right amount of resources to request for a computing job. Making sure that the application will have enough resources to run properly, but avoiding over-requests that would make the jobs spend too much
timestamp1677204000000SRCF is expandingby Kilian Cavalotti, Technical Lead & Architect, HPCMaintenanceIn order to bring up a new building that will increase data center capacity, a full SRCF power shutdown is planned for late June 2023. It’s expected to last about a week, and Sherlock will be unavailable during that time.
timestamp1670036242756ClusterShell on Sherlockby Kilian Cavalotti, Technical Lead & Architect, HPCSoftwareNewEver wondered how your jobs were doing while they were running? Keeping a eye on a log file is nice, but what if you could quickly gather process lists, usage metrics and other data points from all the nodes your multi-node jobs are running
timestamp1667700685989Job #1, again!by Kilian Cavalotti, Technical Lead & Architect, HPCThis is not the first time, we’ve been through this already (not so long ago, actually) but today, the Slurm job id counter was reset and went from job #67043327 back to job #1.