I’ve successfully created the Gravity tarfile for the cluster deployment, and have installed the cluster on a fresh Debian 9 image in GCP with:
sudo ./gravity install --advertise-addr=10.168.0.10 --token=secret
(I confirm I can
gravity enter and then
kubectl get pods shows the active pods).
But I then spin up an identically configured GCP instance on the same VPC, in the same region/AZ, with firewall rules enabling full networking between those instances. And I attempt to join the existing cluster with:
sudo ./gravity join 10.168.0.10 --advertise-addr=10.168.0.11 --token=secret
However, this times out or fails:
Wed Aug 7 17:44:27 UTC Still waiting for the planet to start (8 minutes elapsed)
Wed Aug 7 17:44:27 UTC Saving debug report to /home/user/mattermost/crashreport.tgz
And, on the second host, after
sudo gravity enter, I attempt to diagnose with
kubectl get -o pods but that eventually produces:
Unable to connect to the server: dial tcp 10.100.0.1:443: i/o timeout
Within the second host’s environment (after I run
sudo gravity enter, that is), there is no indication of any running containers via
docker ps -a.
journalctl -xe gives:
kube-proxy: E0807 18:59:57.805298 109 proxier.go:1402] Failed to execute iptables-restore: exit status 2 (iptables-restore v1.6.0: Couldn't load targetKUBE-MARK-DROP’:No such file or directory
, and thenkube-proxy: Error occurred at line: 81`
I’m wondering what could cause all of this. Where would I look for further error messages? Is there a way to invoke a more failsafe or stable instance known to work?