Deploy storage quota functionality uniformly across all AWS and GCP clusters/hubs so that hub administrators have consistent resource management capabilities to prevent users from consuming excessive storage.
Aggregate per-user cloud costs by JupyterHub group so that hub administrators can transparently see cost attribution by organizational team/department/etc. This builds on per-user cost reporting (#6315) and provides cost rollups where users in multiple groups may be double-counted.
Once π£ Per-user cost reporting with Grafana on AWS is done, we will be able to have per user cost reports for hubs on AWS. Next step is to provide a way to aggregate these by group.
Cloud cost monitoring is an important tool for providing hub administrators with operational oversight of usage and budgets with transparency. This allows them to demonstrate value to funders and make informed decisions on impact and cost. Currently, we do not have a clear way to report the per-user costs of running a hub, which makes it difficult for hub administrators to the value and impact on a per-user basis.
Building on the successful work on deploying BinderHub for Project Pythia, we are close to be able support JupyterHub more generally for other communities wanting to use Jetstream2
The missing technical piece is to allow home directory storage quotas on Jetstream2. Once that has been completed, we should be able to add JetStream2 to our product menu.
Unlike AWS and Azure, GCP doesnβt have a way to automatically schedule backups of Filestore. Googleβs own suggested solution is pretty halfbaked, and seems to be focused on βhow many GCP specific features can we throw together?β. The python code they provide to do backups is particularly bad, with no error handling nor retention policies.