Mv pvc k8s

From UVOO Tech Wiki
Revision as of 19:09, 26 February 2025 by Busk (talk | contribs)
Jump to navigation Jump to search

Scripts

Basically create tmp pvc and then copies data from original pvc to tmp pvc, deletes original pvc and then creates new pvc again and copies data back on it.

This is useful for converting disk sizes, storage tier, zones, regions etc. It's a little sloppy now refine in the future

Combined

#!/bin/bash
set -eux

usage() {
  echo "Usage: $0 <old_pvc_name> <new_pvc_size_Gi> <new_storage_class> <namespace>"
  exit 1
}

if [ "$#" -ne 4 ]; then
  usage
fi

OLD_PVC_NAME="$1"
TMP_PVC_NAME="${OLD_PVC_NAME}-tmp"
NEW_PVC_SIZE_GI="$2"
NEW_STORAGE_CLASS="$3"
NAMESPACE="$4"
DATA_COPY_POD_NAME="data-copy-pod"

# Detect the workload using the old PVC and scale it down
POD_NAME=$(kubectl get pods -n "$NAMESPACE" --field-selector=status.phase=Running \
  -o jsonpath="{.items[?(@.spec.volumes[*].persistentVolumeClaim.claimName=='$OLD_PVC_NAME')].metadata.name}" | awk '{print $1}')

if [ -z "$POD_NAME" ]; then
    echo "No running pod found using PVC: $OLD_PVC_NAME"
    exit 1
fi

OWNER_KIND=$(kubectl get pod "$POD_NAME" -n "$NAMESPACE" -o jsonpath="{.metadata.ownerReferences[0].kind}")
OWNER_NAME=$(kubectl get pod "$POD_NAME" -n "$NAMESPACE" -o jsonpath="{.metadata.ownerReferences[0].name}")

case "$OWNER_KIND" in
    "Deployment") kubectl scale deployment "$OWNER_NAME" --replicas=0 -n "$NAMESPACE" ;;
    "StatefulSet") kubectl scale statefulset "$OWNER_NAME" --replicas=0 -n "$NAMESPACE" ;;
    "DaemonSet") kubectl delete daemonset "$OWNER_NAME" -n "$NAMESPACE" ;;
    *) echo "Unsupported workload type: $OWNER_KIND"; exit 1 ;;
esac

echo "Scaled $OWNER_KIND $OWNER_NAME to 0."

# Validate PVC size input
if [[ ! "$NEW_PVC_SIZE_GI" =~ ^[0-9]+$ ]]; then
  echo "Error: New PVC size must be a positive integer (Gi)"
  usage
fi

# Create temporary PVC
kubectl create -f - <<EOF
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: $TMP_PVC_NAME
  namespace: $NAMESPACE
spec:
  accessModes: [ReadWriteOnce] # Or ReadWriteMany if needed
  resources:
    requests:
      storage: ${NEW_PVC_SIZE_GI}Gi
  storageClassName: $NEW_STORAGE_CLASS
EOF

# Create data copy pod
kubectl create -f - <<EOF
apiVersion: v1
kind: Pod
metadata:
  name: $DATA_COPY_POD_NAME
  namespace: $NAMESPACE
spec:
  restartPolicy: Never
  containers:
  - name: data-copy
    image: alpine:latest
    command: ["sh", "-c", "apk add --no-cache rsync && rsync -aHAXv --progress /mnt/old/ /mnt/new/"]
    volumeMounts:
    - name: old-volume
      mountPath: /mnt/old
    - name: new-volume
      mountPath: /mnt/new
  volumes:
  - name: old-volume
    persistentVolumeClaim:
      claimName: $OLD_PVC_NAME
  - name: new-volume
    persistentVolumeClaim:
      claimName: $TMP_PVC_NAME
EOF

# Wait for copy to complete
while true; do
    PHASE=$(kubectl get pod "$DATA_COPY_POD_NAME" -n "$NAMESPACE" -o jsonpath='{.status.phase}')
    if [ "$PHASE" = "Succeeded" ]; then
        echo "Data copy to temporary PVC completed."
        break
    else
        echo "Pod $DATA_COPY_POD_NAME is in phase $PHASE. Waiting 5 seconds..."
        sleep 5
    fi
done

kubectl logs $DATA_COPY_POD_NAME -n $NAMESPACE
kubectl delete pod $DATA_COPY_POD_NAME -n $NAMESPACE
kubectl delete pvc $OLD_PVC_NAME -n $NAMESPACE

# Create new PVC with original name
kubectl create -f - <<EOF
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: $OLD_PVC_NAME
  namespace: $NAMESPACE
spec:
  accessModes: [ReadWriteOnce] # Or ReadWriteMany if needed
  resources:
    requests:
      storage: ${NEW_PVC_SIZE_GI}Gi
  storageClassName: $NEW_STORAGE_CLASS
EOF

# Create data copy pod for final migration
kubectl create -f - <<EOF
apiVersion: v1
kind: Pod
metadata:
  name: $DATA_COPY_POD_NAME
  namespace: $NAMESPACE
spec:
  restartPolicy: Never
  containers:
  - name: data-copy
    image: alpine:latest
    command: ["sh", "-c", "apk add --no-cache rsync && rsync -aHAXv --progress /mnt/old/ /mnt/new/"]
    volumeMounts:
    - name: old-volume
      mountPath: /mnt/old
    - name: new-volume
      mountPath: /mnt/new
  volumes:
  - name: old-volume
    persistentVolumeClaim:
      claimName: $TMP_PVC_NAME
  - name: new-volume
    persistentVolumeClaim:
      claimName: $OLD_PVC_NAME
EOF

# Wait for final copy
while true; do
    PHASE=$(kubectl get pod "$DATA_COPY_POD_NAME" -n "$NAMESPACE" -o jsonpath='{.status.phase}')
    if [ "$PHASE" = "Succeeded" ]; then
        echo "Data copy to new PVC completed."
        break
    else
        echo "Pod $DATA_COPY_POD_NAME is in phase $PHASE. Waiting 5 seconds..."
        sleep 5
    fi
done

kubectl logs $DATA_COPY_POD_NAME -n $NAMESPACE
kubectl delete pod $DATA_COPY_POD_NAME -n $NAMESPACE
kubectl delete pvc $TMP_PVC_NAME -n $NAMESPACE

# Scale workload back up
case "$OWNER_KIND" in
    "Deployment") kubectl scale deployment "$OWNER_NAME" --replicas=1 -n "$NAMESPACE" ;;
    "StatefulSet") kubectl scale statefulset "$OWNER_NAME" --replicas=1 -n "$NAMESPACE" ;;
    *) echo "Skipping scaling for $OWNER_KIND";;
esac

echo "PVC migration completed successfully."

Split

main.sh

#!/bin/bash
set -eu
# kubectl delete ns zabbix
# velero restore create --from-backup zabbix-tobucket --include-namespaces zabbix
# sleep 120
kubectl scale sts/zabbix-postgresql --replicas=0
./mv-pvc.sh postgresql-data-zabbix-postgresql-0 postgresql-data-zabbix-postgresql-0-tmp 7 default zabbix
kubectl delete pvc postgresql-data-zabbix-postgresql-0
./mv-pvc.sh postgresql-data-zabbix-postgresql-0-tmp postgresql-data-zabbix-postgresql-0 7 default zabbix
kubectl scale sts/zabbix-postgresql --replicas=1

recreate-pvc-with-original-data.sh

#!/bin/bash
set -eux

# Function to display usage instructions
usage() {
  echo "Usage: $0 <old_pvc_name> <new_pvc_name> <new_pvc_size_Gi> <new_storage_class> <namespace>"
  exit 1
}

# Check for correct number of arguments
if [ "$#" -ne 5 ]; then
  usage
fi

check_pod_status(){
POD_NAME=$1
while true; do
    PHASE=$(kubectl get pod "$POD_NAME" -n "$NAMESPACE" -o jsonpath='{.status.phase}')
    if [ "$PHASE" = "Succeeded" ]; then
        echo "Pod $POD_NAME has succeeded."
        break
    else
        echo "Pod $POD_NAME is in phase $PHASE. Waiting 5 seconds..."
        sleep 5
    fi
done
}

# Assign arguments to variables
OLD_PVC_NAME="$1"
NEW_PVC_NAME="$2"
NEW_PVC_SIZE_GI="$3"
NEW_STORAGE_CLASS="$4"
NAMESPACE="$5"

# Validate PVC size
if [[ ! "$NEW_PVC_SIZE_GI" =~ ^[0-9]+$ ]]; then
  echo "Error: New PVC size must be a positive integer (Gi)"
  usage
fi


DATA_COPY_POD_NAME="data-copy-pod"

# 1. Create the new PVC
kubectl create -f - <<EOF
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: $NEW_PVC_NAME
  namespace: $NAMESPACE
spec:
  accessModes: [ReadWriteOnce] # Or ReadWriteMany if needed
  resources:
    requests:
      storage: ${NEW_PVC_SIZE_GI}Gi
  storageClassName: $NEW_STORAGE_CLASS
EOF

# 2. Create the data copy pod
kubectl create -f - <<EOF
apiVersion: v1
kind: Pod
metadata:
  name: $DATA_COPY_POD_NAME
  namespace: $NAMESPACE
spec:
  restartPolicy: Never # Important: Only run once
  containers:
  - name: data-copy
    image: alpine:latest
    # command: ["sh", "-c", "apk add --no-cache rsync && rsync -av /mnt/old/ /mnt/new/"]
    command: ["sh", "-c", "apk add --no-cache rsync && rsync -aHAXv --progress /mnt/old/ /mnt/new/"]
    volumeMounts:
    - name: old-volume
      mountPath: /mnt/old
    - name: new-volume
      mountPath: /mnt/new
  volumes:
  - name: old-volume
    persistentVolumeClaim:
      claimName: $OLD_PVC_NAME
  - name: new-volume
    persistentVolumeClaim:
      claimName: $NEW_PVC_NAME
EOF

# 3. Wait for the data copy to finish (adjust timeout as needed)
# kubectl wait --for=condition=Succeeded --timeout=5m pod/$DATA_COPY_POD_NAME -n $NAMESPACE
# kubectl wait --for=condition=Completed --timeout=5m pod/$DATA_COPY_POD_NAME -n $NAMESPACE
check_pod_status ${DATA_COPY_POD_NAME}

# 4. Check the status of the copy pod
kubectl logs $DATA_COPY_POD_NAME -n $NAMESPACE
kubectl get pod $DATA_COPY_POD_NAME -n $NAMESPACE

# 5. (Optional) Update your application deployment to use the new PVC

# 6. (After verifying everything is working) Delete the data copy pod
kubectl delete pod $DATA_COPY_POD_NAME -n $NAMESPACE

# 7. Delete the old PVC (after you're *absolutely sure* you don't need it)
# kubectl delete pvc $OLD_PVC_NAME -n $NAMESPACE

# 8. Delete the old PV (if reclaimPolicy was Retain)
# kubectl get pv # Find the PV associated with the old PVC
# kubectl delete pv <pv-name>
# command: ["/bin/bash", "-c", "sleep infinity"]

PVC release

#!/bin/bash
set -eux

if [ "$#" -ne 3 ]; then
    echo "Usage: $0 <old-storage-class> <new-storage-class> <namespace|'all'>"
    exit 1
fi

OLD_SC=$1
NEW_SC=$2
NS_FILTER=$3

if [ "$NS_FILTER" == "all" ]; then
    PVC_LIST=$(kubectl get pvc -A -o jsonpath="{range .items[?(@.spec.storageClassName=='$OLD_SC')]}{.metadata.namespace} {.metadata.name}{'\n'}{end}")
else
    PVC_LIST=$(kubectl get pvc -n "$NS_FILTER" -o jsonpath="{range .items[?(@.spec.storageClassName=='$OLD_SC')]}{.metadata.namespace} {.metadata.name}{'\n'}{end}")
fi

echo "$PVC_LIST" | while read -r NS PVC; do
    PV=$(kubectl get pvc "$PVC" -n "$NS" -o jsonpath="{.spec.volumeName}")
    STORAGE_SIZE=$(kubectl get pvc "$PVC" -n "$NS" -o jsonpath="{.spec.resources.requests.storage}")

    kubectl patch pv "$PV" --type='json' -p='[{"op": "replace", "path": "/spec/persistentVolumeReclaimPolicy", "value": "Retain"}]'

    kubectl patch pvc "$PVC" --type=merge -p '{"metadata":{"finalizers": []}}' || true
    kubectl delete pvc "$PVC" -n "$NS" --wait=true

    kubectl patch pv "$PV" --type='json' -p="[{'op': 'replace', 'path': '/spec/storageClassName', 'value':'$NEW_SC'}]"
    kubectl patch pv "$PV" --type=merge -p '{"spec":{"claimRef": null}}'

    kubectl apply -f - <<EOF
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
  name: $PVC
  namespace: $NS
spec:
  storageClassName: $NEW_SC
  accessModes:
    - ReadWriteOnce
  resources:
    requests:
      storage: $STORAGE_SIZE
  volumeName: $PV
EOF

    # Restore original reclaim policy (optional)
    kubectl patch pv "$PV" --type='json' -p='[{"op": "replace", "path": "/spec/persistentVolumeReclaimPolicy", "value": "Delete"}]'
done

WIP

#!/bin/bash

# Script to take a snapshot of an AKS disk and create a new disk from it in a different zone, then rename it back to the original name.

# Usage: ./snapshot_and_clone_rename.sh <resource_group> <disk_name> <source_zone> <target_zone>

# Check for required arguments
if [ $# -ne 4 ]; then
  echo "Usage: $0 <resource_group> <disk_name> <source_zone> <target_zone>"
  exit 1
fi

RESOURCE_GROUP="$1"
DISK_NAME="$2"
SOURCE_ZONE="$3"
TARGET_ZONE="$4"
SNAPSHOT_NAME="${DISK_NAME}-snapshot-$(date +%Y%m%d%H%M%S)"
NEW_DISK_NAME="${DISK_NAME}-temp-cloned-${TARGET_ZONE}" # Temporary name for the cloned disk

# Check if Azure CLI is installed
if ! command -v az &> /dev/null; then
  echo "Azure CLI is not installed. Please install it."
  exit 1
fi

# Check if logged in
if ! az account show &> /dev/null; then
  echo "Please log in to Azure CLI using 'az login'."
  exit 1
fi

# Check if the disk exists
if ! az disk show --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" &> /dev/null; then
  echo "Disk '$DISK_NAME' not found in resource group '$RESOURCE_GROUP'."
  exit 1
fi

# Check if the source zone is valid
if [[ -z "$(az disk show --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --query location -o tsv)" ]]; then
    echo "Disk '$DISK_NAME' does not appear to be zonal. Please ensure that the disk is in a specific zone."
    exit 1
fi

# Check if the target zone is valid
if [[ "$SOURCE_ZONE" == "$TARGET_ZONE" ]]; then
    echo "Source and target zones cannot be the same."
    exit 1
fi

# Get the disk ID
DISK_ID=$(az disk show --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --query id -o tsv)

# Create the snapshot
echo "Creating snapshot '$SNAPSHOT_NAME'..."
az snapshot create \
  --resource-group "$RESOURCE_GROUP" \
  --name "$SNAPSHOT_NAME" \
  --source "$DISK_ID" \
  --location "$(az disk show --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --query location -o tsv)" \
  --zone "$SOURCE_ZONE"

if [ $? -ne 0 ]; then
  echo "Failed to create snapshot."
  exit 1
fi

# Get the snapshot ID
SNAPSHOT_ID=$(az snapshot show --resource-group "$RESOURCE_GROUP" --name "$SNAPSHOT_NAME" --query id -o tsv)

# Create the new disk from the snapshot in the target zone with a temporary name
echo "Creating new disk '$NEW_DISK_NAME' in zone '$TARGET_ZONE'..."
az disk create \
  --resource-group "$RESOURCE_GROUP" \
  --name "$NEW_DISK_NAME" \
  --source "$SNAPSHOT_ID" \
  --location "$(az disk show --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --query location -o tsv)" \
  --zone "$TARGET_ZONE"

if [ $? -ne 0 ]; then
  echo "Failed to create new disk."
  # Optionally delete the snapshot if disk creation fails
  az snapshot delete --resource-group "$RESOURCE_GROUP" --name "$SNAPSHOT_NAME"
  exit 1
fi

# Delete the original disk
echo "Deleting original disk '$DISK_NAME'..."
az disk delete --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --yes

if [ $? -ne 0 ]; then
  echo "Failed to delete original disk."
  # Optionally delete the snapshot and new disk if original disk deletion fails
  az snapshot delete --resource-group "$RESOURCE_GROUP" --name "$SNAPSHOT_NAME"
  az disk delete --resource-group "$RESOURCE_GROUP" --name "$NEW_DISK_NAME" --yes
  exit 1
fi

# Rename the new disk to the original name
echo "Renaming new disk '$NEW_DISK_NAME' to '$DISK_NAME'..."
az disk update --resource-group "$RESOURCE_GROUP" --name "$NEW_DISK_NAME" --set name="$DISK_NAME"

if [ $? -ne 0 ]; then
  echo "Failed to rename new disk."
  # Optionally delete the snapshot and new disk if rename fails
  az snapshot delete --resource-group "$RESOURCE_GROUP" --name "$SNAPSHOT_NAME"
  az disk delete --resource-group "$RESOURCE_GROUP" --name "$DISK_NAME" --yes
  exit 1
fi

# Clean up the snapshot (optional)
echo "Deleting snapshot '$SNAPSHOT_NAME'..."
az snapshot delete --resource-group "$RESOURCE_GROUP" --name "$SNAPSHOT_NAME"

echo "Disk '$DISK_NAME' created successfully in zone '$TARGET_ZONE'."

exit 0

Key Changes:
 * Temporary Disk Name:
   * The cloned disk is now created with a temporary name (e.g., myAKSDisk-temp-cloned-westus2). This avoids a name collision with the original disk.
 * Original Disk Deletion:
   * The script now deletes the original disk after the cloned disk is successfully created.
   * Added --yes flag to avoid prompt.
 * Renaming the Cloned Disk:
   * The cloned disk is then renamed to the original disk's name.
   * Uses az disk update --set name to rename the disk.
 * Enhanced Error Handling:
   * Added error checking for the original disk deletion and rename operation.
   * Added cleanup routines in case of deletion or rename failures.
Important Considerations:
 * Data Loss Risk: Be extremely careful when using this script, as it involves deleting the original disk. Make sure you have backups or understand the implications before running it.
 * Downtime: This process will cause downtime for any services that rely on the disk.
 * Testing: Thoroughly test this script in a non-production environment before using it in production.
 * AKS Integration: If the disk is used by an AKS node pool, you will likely need to perform additional steps to update the node pool configuration after the disk is moved. Consider using node pool scaling to replace the nodes using the old disks with the new disks.