I believe I’ve found an interesting race condition during EC2 instance creation due to a slow-running cloud-init process. The issue is that cloud-init appears to create the initial login user & installs the public SSH key onto a newly started EC2 instance, then restarts sshd. It takes a while to do this, and creates a race condition where Ansible cannot connect to the host and fails the playbook run. In my playbook, I’m using the ec2 module, followed by add_host, and then wait_for to wait for the SSH port to be open. I have also experimented with using a simple “shell: echo host_is_up” command with a retry / do-until loop. However this also fails because Ansible wants the initial SSH connection to be successful, which it will not in this case. So Ansible does not retry
It appears that due to the user not existing until ~3 minutes after it is booted and sshd is listening on port 22, Ansible cannot connect as the initial login user for the CentOS AMI (“centos”). So the SSH port open check is not good enough to detect and wait for the port to be open AND the login user to exist. The simple echo shell command with retry do/until loop also does not work, because the very first SSH connection Ansible tries to make to run the module fails also.
My question is: Has anyone run into a similar issue with EC2 instances being slow to become available causing Ansible to fail to connect, and also found a solution to this?
I realize that a sleep task is one possible solution (and I may be forced to reach for that sledgehammer), but it doesn’t feel like the absolute best solution because we really want to wait for both cloud-init to be finished creating “centos” user on the instance AND SSH to be up. So really, the only other way I can think of is to somehow tell SSH to retry connecting as centos until it succeeds or a surpasses a very long timeout. Is this possible? Are there better ways of handling this?
I just discovered this issue as well, with various random ssh connection or generic authentication/permission failure messages, some occurring after a play successfully passed a task or two. It occurred very consistently with many CentOS 7 t2.nano hosts. A 2-minute pause after waiting for ssh listener resolved it for me. The system logs showed the centos user added about 1 minute into boot time, so I gave it two minutes to be generous:
`
name: wait for instances to listen on port:22
wait_for:
state: started
host: “{{ item }}”
port: 22
name: wait for boot process to finish
pause: minutes=2
`
It also helped to make sure I removed old host keys, even though I have strict checking turned off:
PS) Some of the errors I saw caused by this:
“failed to transfer file to /tmp/.ansible”
“Authentication or permission failure. In some cases, you may have been able to authenticate and did not have permissions on the remote directory. Consider changing the remote temp path in ansible.cfg to a path rooted in "/tmp". Failed command was: ( umask 77 && mkdir -p "` echo /tmp/.ansible/…”
In that case he’s using uri with an ‘until’ and using the ‘default’ filter to make sure no results doesn’t mean a task failure.
- name: run test uri: url: "[https://0.0.0.0:3030/api/canary](https://0.0.0.0:3030/api/canary)" validate_certs: no register: result until: result['status']|default(0) == 200
Obviously you’d have to replace uri with something that tells you the instance is ready to start working, but it might give you a way to get rid of the need for the arbitrary pause.