Troubleshooting your DC/OS installation

During DC/OS installation, each of the components will converge from a failing state to a running state in the logs.

Take a look at the boot sequence to understand what order everything comes up in.

ZooKeeper and Exhibitor

ZooKeeper and Exhibitor start on the master nodes. The Exhibitor storage location must be configured properly for this to work. For more information, see the exhibitor_storage_backend parameter.

DC/OS uses ZooKeeper, a high-performance coordination service to manage the installed DC/OS services. Exhibitor automatically configures ZooKeeper on the master nodes during your DC/OS installation. For more information, see Configuration Parameters.

Go to the Exhibitor web interface and view status at /exhibitor. SSH to your master node and enter this command to view the logs from boot time:

$ journalctl -u dcos-exhibitor -b For example, here is a snippet of the Exhibitor log as it converges to a successful state:

INFO com.netflix.exhibitor.core.activity.ActivityLog Automatic Instance Management will change the server list: ==> 1:10.0.7.166 [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog State: serving [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog Server list has changed [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog Attempting to stop instance [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog Attempting to start/restart ZooKeeper [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog Kill attempted result: 0 [ActivityQueue-0] INFO com.netflix.exhibitor.core.activity.ActivityLog Process started via: /opt/mesosphere/active/exhibitor/usr/zookeeper/bin/zkServer.sh [ActivityQueue-0] ERROR com.netflix.exhibitor.core.activity.ActivityLog ZooKeeper Server: JMX enabled by default [pool-3-thread-1] ERROR com.netflix.exhibitor.core.activity.ActivityLog ZooKeeper Server: Using config: /opt/mesosphere/active/exhibitor/usr/zookeeper/bin/../conf/zoo.cfg [pool-3-thread-1] INFO com.netflix.exhibitor.core.activity.ActivityLog ZooKeeper Server: Starting zookeeper ... STARTED [pool-3-thread-3] INFO com.netflix.exhibitor.core.activity.ActivityLog Cleanup task completed [pool-3-thread-6] INFO com.netflix.exhibitor.core.activity.ActivityLog Cleanup task completed [pool-3-thread-9] Mesos master process

The Mesos master process starts on the master nodes. The mesos-master process runs on a node in the cluster and orchestrates the running of tasks on agents by receiving resource offers from agents and offering those resources to registered services, such as Marathon or Chronos. For more information, see the Mesos Master Configuration documentation.

Troubleshooting:

Go directly to the Mesos web interface and view status at /mesos. SSH to your master node and enter this command to view the logs from boot time: $ journalctl -u dcos-mesos-master -b For example, here is a snippet of the Mesos master log as it converges to a successful state:

mesos-master[1250]: I1118 13:59:33.890916 1250 master.cpp:376] Master cdcb6222-65a1-4d60-83af-33dadec41e92 (10.0.7.166) started on 10.0.7.166:5050 mesos-master[1250]: I1118 13:59:33.890945 1250 master.cpp:378] Flags at startup: --allocation_interval="1secs" --allocator="HierarchicalDRF" --authenticate="false" --authenticate_slaves="false" --authenticators="crammd5" --authorizers="local" --cluster="pool-880dfdbf0f2845bf8191" --framework_sorter="drf" --help="false" --hostname_lookup="false" --initialize driver_logging="true" --ip_discovery_command="/opt/mesosphere/bin/detect_ip" --log_auto_initialize="true" --log_dir="/var/log/mesos" --logbufsecs="0" --logging_level="INFO" --max slave_ping_timeouts="5" --port="5050" --quiet="false" --quorum="1" --recovery_slave_removal_limit="100%" --registry="replicated_log" --registry_fetch_timeout="1mins" --registry_sto re_timeout="5secs" --registry_strict="false" --roles="slave_public" --root_submissions="true" --slave_ping_timeout="15secs" --slave_reregister_timeout="10mins" --user_sorter="drf" --version="false" --webui_dir="/opt/mesosphere/packages/mesos--30d3fbeb6747bb086d71385e3e2e0eb74ccdcb8b/share/mesos/webui" --weights="slave_public=1" --work_dir="/var/lib/mesos/mas ter" --zk="zk://127.0.0.1:2181/mesos" --zk_session_timeout="10secs" mesos-master[1250]: 2015-11-18 13:59:33,891:1250(0x7f14427fc700):ZOO_INFO@check_events@1750: session establishment complete on server [127.0.0.1:2181], sessionId=0x1511ae440bc0001, negotiated timeout=10000 Mesos-DNS

Mesos-DNS is started on the DC/OS master nodes. Mesos-DNS provides service discovery within the cluster. Optionally, Mesos-DNS can forward unhandled requests to an external DNS server, depending on how the cluster is configured. For example, anything that does not end in .mesos will be forwarded to the external resolver.

Troubleshooting:

SSH to your master node and enter this command to view the logs from boot time: $ journalctl -u dcos-mesos-dns -b For example, here is a snippet of the Mesos-DNS log as it converges to a successful state:

mesos-dns[1197]: I1118 13:59:34.763885 1197 detect.go:135] changing leader node from "" -> "json.info_0000000001" mesos-dns[1197]: I1118 13:59:34.764537 1197 detect.go:145] detected master info: &MasterInfo{Id:cdcb6222-65a1-4d60-83af-33dadec41e92,Ip:2785476618,Port:5050,Pid:master@10.0.7.166:5050,Hostname:10.0.7.166,Version:0.25.0,Address:&Address{Hostname:10.0.7.166,Ip:10.0.7.166,Port:5050,XXX_unrecognized:[],},XXX_unrecognized:[],} mesos-dns[1197]: VERY VERBOSE: 2015/11/18 13:59:34 masters.go:47: Updated leader: &MasterInfo{Id:cdcb6222-65a1-4d60-83af-33dadec41e92,Ip:2785476618,Port:5050,Pid:master@10.0.7.166:5050,Hostname:10.0.7.166,Version:0.25.0,Address:&Address{Hostname:10.0.7.166,Ip:10.0.7.166,Port:5050,XXX_unrecognized:[],},XXX_unrecognized:[],} mesos-dns[1197]: VERY VERBOSE: 2015/11/18 13:59:34 main.go:76: new masters detected: [10.0.7.166:5050] mesos-dns[1197]: VERY VERBOSE: 2015/11/18 13:59:34 generator.go:70: Zookeeper says the leader is: 10.0.7.166:5050 mesos-dns[1197]: VERY VERBOSE: 2015/11/18 13:59:34 generator.go:162: reloading from master 10.0.7.166 mesos-dns[1197]: I1118 13:59:34.766005 1197 detect.go:219] notifying of master membership change: [&MasterInfo{Id:cdcb6222-65a1-4d60-83af-33dadec41e92,Ip:2785476618,Port:5050,Pid:master@10.0.7.166:5050,Hostname:10.0.7.166,Version:0.25.0,Address:&Address{Hostname:10.0.7.166,Ip:10.0.7.166,Port:5050,XXX_unrecognized:[],},XXX_unrecognized:[],}] mesos-dns[1197]: VERY VERBOSE: 2015/11/18 13:59:34 masters.go:56: Updated masters: [&MasterInfo{Id:cdcb6222-65a1-4d60-83af-33dadec41e92,Ip:2785476618,Port:5050,Pid:master@10.0.7.166:5050,Hostname:10.0.7.166,Version:0.25.0,Address:&Address{Hostname:10.0.7.166,Ip:10.0.7.166,Port:5050,XXX_unrecognized:[],},XXX_unrecognized:[],}] mesos-dns[1197]: I1118 13:59:34.766124 1197 detect.go:313] resting before next detection cycle DC/OS Marathon

DC/OS Marathon is started on the master nodes. The native Marathon instance that is the “init system” for DC/OS. It starts and monitors applications and services.

Troubleshooting:

Go to the Services tab on the web interface and view status. SSH to your master node and enter this command to view the logs from boot time:

$ journalctl -u dcos-marathon -b For example, here is a snippet of the DC/OS Marathon log as it converges to a successful state:

java[1288]: I1118 13:59:39.125041 1363 group.cpp:331] Group process (group(1)@10.0.7.166:48531) connected to ZooKeeper java[1288]: I1118 13:59:39.125100 1363 group.cpp:805] Syncing group operations: queue size (joins, cancels, datas) = (0, 0, 0) java[1288]: I1118 13:59:39.125121 1363 group.cpp:403] Trying to create path '/mesos' in ZooKeeper java[1288]: [2015-11-18 13:59:39,130] INFO Scheduler actor ready (mesosphere.marathon.MarathonSchedulerActor:marathon-akka.actor.default-dispatcher-5) java[1288]: I1118 13:59:39.147804 1363 detector.cpp:156] Detected a new leader: (id='1') java[1288]: I1118 13:59:39.147924 1363 group.cpp:674] Trying to get '/mesos/json.info_0000000001' in ZooKeeper java[1288]: I1118 13:59:39.148727 1363 detector.cpp:481] A new leading master (UPID=master@10.0.7.166:5050) is detected java[1288]: I1118 13:59:39.148787 1363 sched.cpp:262] New master detected at master@10.0.7.166:5050 java[1288]: I1118 13:59:39.148952 1363 sched.cpp:272] No credentials provided. Attempting to register without authentication java[1288]: I1118 13:59:39.150403 1363 sched.cpp:641] Framework registered with cdcb6222-65a1-4d60-83af-33dadec41e92-0000 Admin Router

The Admin Router is started on the master nodes. The Admin Router provides central authentication and proxy to DC/OS services within the cluster. This allows you to administer your cluster from outside the network without VPN or a SSH tunnel. For HA, an optional load balancer can be configured in front of each master node, load balancing port 80, to provide failover and load balancing.

Troubleshooting:

SSH to your master node and enter this command to view the logs from boot time: $ journalctl -u dcos-adminrouter -b For example, here is a snippet of the Admin Router log as it converges to a successful state:

systemd[1]: Starting A high performance web server and a reverse proxy server... systemd[1]: Started A high performance web server and a reverse proxy server. nginx[1652]: ip-10-0-7-166.us-west-2.compute.internal nginx: 10.0.7.166 - - [18/Nov/2015:14:01:10 +0000] "GET /mesos/master/state-summary HTTP/1.1" 200 575 "-" "python-requests/2.6.0 CPython/3.4.2 Linux/4.1.7-coreos" nginx[1652]: ip-10-0-7-166.us-west-2.compute.internal nginx: 10.0.7.166 - - [18/Nov/2015:14:01:10 +0000] "GET /metadata HTTP/1.1" 200 175 "-" "python-requests/2.6.0 CPython/3.4.2 Linux/4.1.7-coreos" gen_resolvconf

gen_resolvconf is started. This is a service that helps the agent nodes locate the master nodes. It updates /etc/resolv.conf so that agents can use the Mesos-DNS service for service discovery. gen_resolvconf uses either an internal load balancer, vrrp, or a static list of masters to locate the master nodes. For more information, see the master_discovery configuration parameter.

Troubleshooting:

When gen_resolvconf is up and running, you can view /etc/resolv.conf contents. It should contain one or more IP addresses for the master nodes, and the optional external DNS server. SSH to your master node and enter this command to view the logs from boot time:

$ journalctl -u dcos-gen-resolvconf -b For example, here is a snippet of the gen_resolvconf log as it converges to a successful state:

systemd[1]: Started Update systemd-resolved for mesos-dns. systemd[1]: Starting Update systemd-resolved for mesos-dns... gen_resolvconf.py[1073]: options timeout:1 gen_resolvconf.py[1073]: options attempts:3 gen_resolvconf.py[1073]: nameserver 10.0.7.166 gen_resolvconf.py[1073]: nameserver 10.0.0.2 gen_resolvconf.py[1073]: Updating /etc/resolv.conf DC/OS agent nodes

DC/OS private and public agent nodes are started. Deployed apps and services are run on the private agent nodes. You must have at least 1 private agent node.

Publicly accessible applications are run in the public agent node. Public agent nodes can be configured to allow outside traffic to access your cluster. Public agents are optional and there is no minimum. This is where you’d run a load balancer, providing a service from inside the cluster to the external public.

Troubleshooting:

You might not be able to SSH to agent nodes, depending on your cluster network configuration. We have made this a little bit easier with the DC/OS CLI. For more information, see SSHing to a DC/OS cluster. You can get the IP address of registered agent nodes from the Nodes tab in the DC/OS web interface. Nodes that have not registered are not shown.

SSH to your agent node and enter this command to view the logs from boot time:

$ journalctl -u dcos-mesos-slave -b For example, here is a snippet of the Mesos agent log as it converges to a successful state:

mesos-slave[1080]: I1118 14:00:43.687366 1080 main.cpp:272] Starting Mesos slave mesos-slave[1080]: I1118 14:00:43.688474 1080 slave.cpp:190] Slave started on 1)@10.0.1.108:5051 mesos-slave[1080]: I1118 14:00:43.688503 1080 slave.cpp:191] Flags at startup: --appc_store_dir="/tmp/mesos/store/appc" --authenticatee="crammd5" --cgroups_cpu_enable_pids_and_tids_count="false" --cgroups_enable_cfs="false" --cgroups_hierarchy="/sys/fs/cgroup" --cgroups_limit_swap="false" --cgroups_root="mesos" --container_disk_watch_interval="15secs" --containerizers="docker,mesos" --default_role="" --disk_watch_interval="1mins" --docker="docker" --docker_kill_orphans="true" --docker_remove_delay="1hrs" --docker_socket="/var/run/docker.sock" --docker_stop_timeout="0ns" --enforce_container_disk_quota="false" --executor_environment_variables="{"LD_LIBRARY_PATH":"\/opt\/mesosphere\/lib","PATH":"\/usr\/bin","SASL_PATH":"\/opt\/mesosphere\/lib\/sasl2","SHELL":"\/usr\/bin\/bash"}" --executor_registration_timeout="5mins" --executor_shutdown_grace_period="5secs" --fetcher_cache_dir="/tmp/mesos/fetch" --fetcher_cache_size="2GB" --frameworks_home="" --gc_delay="2days" --gc_disk_headroom="0.1" --hadoop_home="" --help="false" --hostname_lookup="false" --image_provisioner_backend="copy" --initialize_driver_logging="true" --ip_discovery_command="/opt/mesosphere/bin/detect_ip" --isolation="cgroups/cpu,cgroups/mem" --launcher_dir="/opt/mesosphere/packages/mesos--30d3fbeb6747bb086d71385e3e2e0eb74ccdcb8b/libexec/mesos" --log_dir="/var/log/mesos" --logbufsecs="0" --logging_level="INFO" --master="zk://leader.mesos:2181/mesos" --oversubscribed_resources_interval="15secs" --perf_duration="10secs" --perf_interval="1mins" --port="5051" --qos_correction_interval_min="0ns" --quiet="false" --recover="reconnect" --recovery_timeout="15mins" --registration_backoff_factor="1secs" --resource_monitoring_interval="1secs" --resources="ports:[1025-2180,2182-3887,3889-5049,5052-8079,8082-8180,8182-32000]" --revocable_cpu_low_priority="true" --sandbox_directory="/mnt/mesos/sandbox" --slave_subsystems="cpu,memory" --strict="true" --switch_user="true" --systemd_runtime_directory="/run/systemd/system" --version="false" --work_dir="/var/lib/mesos/slave" mesos-slave[1080]: I1118 14:00:43.688711 1080 slave.cpp:211] Moving slave process into its own cgroup for subsystem: cpu mesos-slave[1080]: 2015-11-18 14:00:43,689:1080(0x7f9b526c4700):ZOO_INFO@check_events@1703: initiated connection to server [10.0.7.166:2181] mesos-slave[1080]: I1118 14:00:43.692811 1080 slave.cpp:211] Moving slave process into its own cgroup for subsystem: memory mesos-slave[1080]: I1118 14:00:43.697872 1080 slave.cpp:354] Slave resources: ports():[1025-2180, 2182-3887, 3889-5049, 5052-8079, 8082-8180, 8182-32000]; cpus():4; mem():14019; disk(*):32541 mesos-slave[1080]: I1118 14:00:43.697916 1080 slave.cpp:390] Slave hostname: 10.0.1.108 mesos-slave[1080]: I1118 14:00:43.697928 1080 slave.cpp:395] Slave checkpoin

Stop waiting task and restart it

dcos marathon task list
dcos marathon task show spark4ad.f852095b-dbec-11e7-a9d1-92c158367ffc
dcos marathon task stop --wipe spark4ad.f852095b-dbec-11e7-a9d1-92c158367ffc

Last updated

Was this helpful?