I have been playing with Rancher/K3S on a 3 nodes cluster with 1 master and two agents. I have been testing the backup restore process before running any workloads. In doing the restore I have been banging my head on my desk trying to figure out why
freenas.gravyflex.ca is resolving to my router’s public address, instead of the local address. I have verified that DNS resolves to the local address. Any ideas on what could be causing this?
ERRO[2022/05/04 21:29:30] error syncing 'restore-migration': handler restore: failed to check s3 bucket:k3s, err:Get "https://freenas.gravyflex.ca:9000/k3s/?location=": dial tcp <PUBLIC IP>:9000: i/o timeout, requeuing INFO[2022/05/04 21:29:30] Processing Restore CR restore-migration INFO[2022/05/04 21:29:30] Restoring from backup may3-e095543e-f4db-4472-a114-557927fef785-2022-05-03T23-36-38Z.tar.gz INFO[2022/05/04 21:29:30] invoking set s3 service client s3-accessKey=<REDACTED> s3-bucketName=k3s s3-endpoint="freenas.gravyflex.ca:9000" s3-endpoint-ca= s3-folder= s3-region=
# migrationResource.yaml apiVersion: resources.cattle.io/v1 kind: Restore metadata: name: restore-migration spec: backupFilename: may3-e095543e-f4db-4472-a114-557927fef785-2022-05-03T23-36-38Z.tar.gz prune: false storageLocation: s3: credentialSecretName: creds credentialSecretNamespace: default bucketName: k3s endpoint: freenas.gravyflex.ca:9000