I am attempting to restrict access to servers in my farm and as a test
attempted the following with just one instance.
I first changed my "default" group to include the
following:
user: 919814621061
group: scalr-ssh2
and
user: 919814621061
group: scalr-snmp
I was assuming this would make sure that Scalr could access ssh and
snmp even if I block it from the security group. However upon doing
this Scalr immediately loses SNMP contact with the instance, and a
"Synchronize to All" shows that SSH is blocked as well because I got
this error:
Apr 3 22:14:45 ec2-72-44-39-234 logger: SCALR(instance-init.sh):
ec2_submit_event(hostInit): Cannot upload keys on 'i-9e3c5cf7'. Failed
to connect to '
72.44.39.234:22'.
So, any ideas from the amazing Scalr devs would be excellent. I don't
mind if I have to put a bunch of static ips into my default security
group for Scalr. I think you guys are hosted at The Planet right? I am
assuming you probably have static ips. However, some kind of fix that
doesn't require ips would be excellent, because it would have to
constantly be updated as you guys added/removed servers etc.
Any ideas?