The Kubernetes cluster will undergo a major hardware update in October.
'''TL;DR: a complete cluster reinstallation will be necessary due to major changes in the underlying network hardware. New persistent storage will be installed, and all persistent volumes will need to be deleted, as the drives will be integrated into the new system. Please start to backup everything and be prepared to delete terminate all your pods and PVs on short notice.'''
Some details in using the cluster after the reinstallation will change slightly. Some of them you can test already, please do so and help me find possible bugs before all changes go live. See below for more information about the major changes.
== Preview: changes in persistent volumes ==
All local NVMe and SSD drives will be integrated into the ceph storage cluster. The nodes will not provide local PVs anymore, and there is only one cluster-wide global storage class. You will finally be able to mount PVs read/write on different pods across different nodes. Special read-only storage for shared datasets will be provided as before, but driven by ceph instead of a an NFS export.
What you can do right now already to get rid of probably most of the data on your PVs is to move all static datasets to the system-wide storage, as described [[CCU:Global dataset storage|here]]. This framework will persist over the cluster reinstall and no data will be lost.
== Preview: changes in permissions and resource use ==
Too prevent hogging of too many resources at the same time, we will probably have to enforce user-specific resource quotas based on their group membership. Also, access to specific more or less powerful nodes based on group memberships will in the future be enforced. Exceptions can always be negotiated, for example in the final phase before a paper submission. It is likely that there also will be an allowed maximum to the lifetime of a pod.
== Preview: changes behind the scenes, new nodes ==