The installation procedure did also install the required gems and after it finished, I again tried to start AppScale and now I get this result:
# head -n150 /var/log/appscale/controller-17443.log
I, [2016-09-19T11:08:14.631387 #15774] INFO -- : ==== Starting AppController (pid: 15774) ====
E, [2016-09-19T11:08:15.820336 #15774] ERROR -- DjinnServer: OpenSSL::SSL::SSLError: SSL_accept SYSCALL returned=5 errno=0 state=unknown state
/usr/lib/ruby/1.9.1/openssl/ssl-internal.rb:172:in `accept'
I, [2016-09-19T11:08:16.125938 #15774] INFO -- : Converting 'verbose' with value 'False'.
I, [2016-09-19T11:08:16.126343 #15774] INFO -- : Converting 'alter_etc_resolv' with value 'False'.
I, [2016-09-19T11:08:16.126551 #15774] INFO -- : Converting 'ips' with value '[]'.
I, [2016-09-19T11:08:16.126747 #15774] INFO -- : Converting 'table' with value 'cassandra'.
I, [2016-09-19T11:08:16.127024 #15774] INFO -- : Converting 'user_commands' with value '[]'.
I, [2016-09-19T11:08:16.127267 #15774] INFO -- : Converting 'autoscale' with value 'True'.
I, [2016-09-19T11:08:16.127459 #15774] INFO -- : Converting 'replication' with value '1'.
I, [2016-09-19T11:08:16.127659 #15774] INFO -- : Converting 'hostname' with value '10.0.2.102'.
I, [2016-09-19T11:08:16.127833 #15774] INFO -- : Converting 'flower_password' with value 'appscale'.
I, [2016-09-19T11:08:16.128005 #15774] INFO -- : Converting 'max_memory' with value '400'.
I, [2016-09-19T11:08:16.128193 #15774] INFO -- : Converting 'clear_datastore' with value 'False'.
I, [2016-09-19T11:08:16.128415 #15774] INFO -- : Converting 'keyname' with value 'appscale790533e13a9f4294b2dfdab2fcc0bd3a'.
I, [2016-09-19T11:08:16.128605 #15774] INFO -- : Converting 'appengine' with value '1'.
W, [2016-09-19T11:08:16.129035 #15774] WARN -- : max_images is less than the number of nodes!
I, [2016-09-19T11:08:16.668074 #15774] INFO -- : Starting iaas_manager on port 17444 with start command [/usr/bin/python2 /root/appscale/InfrastructureManager/infrastructure_manager_service.py] and stop command [/usr/bin/python2 /root/appscale/scripts/stop_service.py /root/appscale/InfrastructureManager/infrastructure_manager_service.py /usr/bin/python2]
I, [2016-09-19T11:08:16.796947 #15774] INFO -- : Started InfrastructureManager successfully!
I, [2016-09-19T11:08:16.797497 #15774] INFO -- : /etc/appscale/zookeeper_locations.json doesn't exist: not restoring data.
I, [2016-09-19T11:08:16.846542 #15774] INFO -- : Reloading nginx service.
[ALERT] 262/110817 (15872) : config : no <listen> line. Nothing to do !
[ALERT] 262/110817 (15872) : Fatal errors found in configuration.
I, [2016-09-19T11:08:17.286290 #15774] INFO -- : Spawning/setting up other nodes.
I, [2016-09-19T11:08:17.287166 #15774] INFO -- : Nodes info after starting remotes: .
I, [2016-09-19T11:08:17.564989 #15774] INFO -- : Done initializing nodes.
I, [2016-09-19T11:08:19.779658 #15774] INFO -- : Starting haproxy on port 9999 with start command [/usr/sbin/service haproxy start] and stop command [/usr/sbin/service haproxy stop]
I, [2016-09-19T11:08:19.892680 #15774] INFO -- : HAProxy configured and started.
I, [2016-09-19T11:08:23.255697 #15774] INFO -- : Starting nginx on port 9999 with start command [/usr/bin/service nginx start] and stop command [/usr/bin/service nginx stop]
monit: action failed -- There is no service by that name
I, [2016-09-19T11:08:23.426713 #15774] INFO -- : Nginx configured and started.
cat: /var/run/haproxy.pid: No such file or directory
I, [2016-09-19T11:08:23.592029 #15774] INFO -- : Reloading nginx service.
I, [2016-09-19T11:08:23.815880 #15774] INFO -- : Reloading nginx service.
I, [2016-09-19T11:08:23.984620 #15774] INFO -- : Only saw one machine, therefore my node is also a slave node
I, [2016-09-19T11:08:27.509696 #15774] INFO -- : Starting API Services.
I, [2016-09-19T11:08:27.511124 #15774] INFO -- : Waiting for DB services ...
I, [2016-09-19T11:08:27.519983 #15774] INFO -- : Starting database services.
I, [2016-09-19T11:08:27.520638 #15774] INFO -- : Starting up Cassandra seed node
I, [2016-09-19T11:08:27.629082 #15774] INFO -- : Starting zookeeper.
[ ? ] console-setup
[ ? ] dns-clean
[ ? ] ejabberd
[ ? ] irqbalance
[ ? ] killprocs
[ ? ] kmod
[ ? ] networking
[ ? ] ondemand
[ ? ] pppd-dns
I, [2016-09-19T11:08:29.674158 #15774] INFO -- : Starting cassandra on port 9999 with start command [/opt/cassandra/cassandra/bin/cassandra start -p /var/appscale/appscale-cassandra.pid] and stop command [/bin/bash -c 'kill -s SIGTERM `cat /var/appscale/appscale-cassandra.pid`']
[ ? ] rc.local
[ ? ] sendsigs
[ ? ] umountfs
[ ? ] umountnfs.sh
[ ? ] umountroot
[ - ] zookeeper
I, [2016-09-19T11:08:34.559338 #15774] INFO -- : Initializing ZooKeeper.
stop: Unknown instance:
I, [2016-09-19T11:08:36.702109 #15774] INFO -- : Starting zookeeper on port 2181 with start command [/usr/sbin/service zookeeper start] and stop command [/usr/sbin/service zookeeper stop]
monit: action failed -- There is no service by that name
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
I, [2016-09-19T11:08:36.823621 #15774] INFO -- : Done configuring zookeeper.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
nodetool: Failed to connect to '
127.0.0.1:7199' - ConnectException: 'Connection refused'.
Error: The node does not have system_traces yet, probably still bootstrapping
error: -1
-- StackTrace --
java.lang.ArrayIndexOutOfBoundsException: -1
at java.util.ArrayList.elementData(ArrayList.java:400)
at java.util.ArrayList.get(ArrayList.java:413)
at org.apache.cassandra.dht.ByteOrderedPartitioner.describeOwnership(ByteOrderedPartitioner.java:255)
at org.apache.cassandra.service.StorageService.effectiveOwnership(StorageService.java:4017)
at org.apache.cassandra.service.StorageService.effectiveOwnership(StorageService.java:87)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.Trampoline.invoke(MethodUtil.java:75)
at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.reflect.misc.MethodUtil.invoke(MethodUtil.java:279)
at com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:112)
at com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:46)
at com.sun.jmx.mbeanserver.MBeanIntrospector.invokeM(MBeanIntrospector.java:237)
at com.sun.jmx.mbeanserver.PerInterface.invoke(PerInterface.java:138)
at com.sun.jmx.mbeanserver.MBeanSupport.invoke(MBeanSupport.java:252)
at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:819)
at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:801)
at javax.management.remote.rmi.RMIConnectionImpl.doOperation(RMIConnectionImpl.java:1488)
at javax.management.remote.rmi.RMIConnectionImpl.access$300(RMIConnectionImpl.java:97)
at javax.management.remote.rmi.RMIConnectionImpl$PrivilegedOperation.run(RMIConnectionImpl.java:1329)
at javax.management.remote.rmi.RMIConnectionImpl.doPrivilegedOperation(RMIConnectionImpl.java:1421)
at javax.management.remote.rmi.RMIConnectionImpl.invoke(RMIConnectionImpl.java:849)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at sun.rmi.server.UnicastServerRef.dispatch(UnicastServerRef.java:323)
at sun.rmi.transport.Transport$2.run(Transport.java:202)
at sun.rmi.transport.Transport$2.run(Transport.java:199)
at java.security.AccessController.doPrivileged(Native Method)
at sun.rmi.transport.Transport.serviceCall(Transport.java:198)
at sun.rmi.transport.tcp.TCPTransport.handleMessages(TCPTransport.java:567)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run0(TCPTransport.java:828)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.access$400(TCPTransport.java:619)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler$1.run(TCPTransport.java:684)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler$1.run(TCPTransport.java:681)
at java.security.AccessController.doPrivileged(Native Method)
at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run(TCPTransport.java:681)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Datacenter: datacenter1
=======================
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
-- Address Load Tokens Owns (effective) Host ID Rack
UN 10.0.2.102 100.7 KB 256 100.0% 61c19e35-94e7-4d0e-94e9-7856c30e3c4c rack1
2016-09-19 11:13:47,189 INFO policies.py:283 Using datacenter 'datacenter1' for DCAwareRoundRobinPolicy (via host '10.0.2.102'); if incorrect, please specify a local_dc to the constructor, or limit contact points to local cluster nodes
... Now, the file gets constantly filled with these error messages ...