This may be a networking issue but ...
I have a master and slave running on EC2 -- the master can see the slave, the slave sees the master and from inside AWS, I can launch a job against the master using its public DNS name (the Spark UI shows the public DNS names for the master and slave).
On my Mac, cruising around the webUI is fine. However, when I try to run the shell connecting to the same Spark master, everything starts up just fine and then hangs "Connecting to master:" (and no job shows up in the webUI). Here's the log:
Initializing interpreter...
13/02/20 14:56:48 INFO server.Server: jetty-7.5.3.v20111011
Creating SparkContext...
13/02/20 14:56:51 INFO slf4j.Slf4jEventHandler: Slf4jEventHandler started
13/02/20 14:56:51 INFO storage.BlockManagerMaster: Registered BlockManagerMaster Actor
13/02/20 14:56:51 INFO storage.MemoryStore: MemoryStore started with capacity 1340.7 MB.
13/02/20 14:56:51 INFO storage.DiskStore: Created local directory at /var/folders/q5/q7jx6k2x3_j9y_wq8455lr1r0000gn/T/spark-local-20130220145651-66ad
13/02/20 14:56:51 INFO network.ConnectionManager: Bound socket to port 63406 with id = ConnectionManagerId(jdonahue-MacBookPro1,63406)
13/02/20 14:56:51 INFO storage.BlockManagerMaster: Trying to register BlockManager
13/02/20 14:56:51 INFO storage.BlockManagerMaster: Registered BlockManager
13/02/20 14:56:51 INFO server.Server: jetty-7.5.3.v20111011
13/02/20 14:56:51 INFO spark.MapOutputTracker: Registered MapOutputTrackerActor actor
13/02/20 14:56:51 INFO spark.HttpFileServer: HTTP File server directory is /var/folders/q5/q7jx6k2x3_j9y_wq8455lr1r0000gn/T/spark-517a1d1d-f4ef-4d94-938d-b21f98e0cf7a
13/02/20 14:56:51 INFO server.Server: jetty-7.5.3.v20111011
13/02/20 14:56:52 INFO io.IoWorker: IoWorker thread 'spray-io-worker-0' started
13/02/20 14:56:52 INFO server.HttpServer: akka://spark/user/BlockManagerHTTPServer started on /
0.0.0.0:63409Spark context available as sc.
Type in expressions to have them evaluated.
Type :help for more information.
When I do exactly the same command after logging in to a running EC2 instance, it works just fine and I see the message that the spark shell is connected and the job is running.
I thought it might be a problem being inside our corporate firewall, but I repeated the experiment outside the firewall and got the same behavior.
Any advice much appreciated ...