Skip to content

Instantly share code, notes, and snippets.

@guicho271828
Created May 27, 2016 04:17
Show Gist options
  • Save guicho271828/67dc5cc6097dc0ea11d6cfc5a8204222 to your computer and use it in GitHub Desktop.
Save guicho271828/67dc5cc6097dc0ea11d6cfc5a8204222 to your computer and use it in GitHub Desktop.
python example
2016-05-27 04:01:55,119 [DEBUG] CloudFormation client initialized with endpoint https://cloudformation.us-west-2.amazonaws.com
2016-05-27 04:01:55,120 [DEBUG] Describing resource MasterServer in stack cfncluster-pepper
2016-05-27 04:01:55,543 [INFO] -----------------------Starting build-----------------------
2016-05-27 04:01:55,734 [DEBUG] Not setting a reboot trigger as scheduling support is not available
2016-05-27 04:01:55,737 [INFO] Running configSets: default
2016-05-27 04:01:55,738 [INFO] Running configSet default
2016-05-27 04:01:55,740 [INFO] Running config deployConfigFiles
2016-05-27 04:01:55,740 [DEBUG] No packages specified
2016-05-27 04:01:55,741 [DEBUG] No groups specified
2016-05-27 04:01:55,741 [DEBUG] No users specified
2016-05-27 04:01:55,741 [DEBUG] No sources specified
2016-05-27 04:01:55,741 [DEBUG] Writing content to /etc/chef/client.rb
2016-05-27 04:01:55,741 [DEBUG] Setting mode for /etc/chef/client.rb to 000644
2016-05-27 04:01:55,741 [DEBUG] Setting owner 0 and group 0 for /etc/chef/client.rb
2016-05-27 04:01:55,742 [DEBUG] Writing content to /tmp/dna.json
2016-05-27 04:01:55,742 [DEBUG] Content will be serialized as a JSON structure
2016-05-27 04:01:55,742 [DEBUG] Setting mode for /tmp/dna.json to 000644
2016-05-27 04:01:55,742 [DEBUG] Setting owner 0 and group 0 for /tmp/dna.json
2016-05-27 04:01:55,742 [DEBUG] Writing content to /tmp/extra.json
2016-05-27 04:01:55,742 [DEBUG] Setting mode for /tmp/extra.json to 000644
2016-05-27 04:01:55,742 [DEBUG] Setting owner 0 and group 0 for /tmp/extra.json
2016-05-27 04:01:55,742 [DEBUG] Running command jq
2016-05-27 04:01:55,743 [DEBUG] No test for command jq
2016-05-27 04:01:55,997 [INFO] Command jq succeeded
2016-05-27 04:01:55,998 [DEBUG] Command jq output:
2016-05-27 04:01:55,998 [DEBUG] Running command mkdir
2016-05-27 04:01:55,998 [DEBUG] No test for command mkdir
2016-05-27 04:01:56,001 [INFO] Command mkdir succeeded
2016-05-27 04:01:56,001 [DEBUG] Command mkdir output:
2016-05-27 04:01:56,001 [DEBUG] Running command touch
2016-05-27 04:01:56,001 [DEBUG] No test for command touch
2016-05-27 04:01:56,003 [INFO] Command touch succeeded
2016-05-27 04:01:56,003 [DEBUG] Command touch output:
2016-05-27 04:01:56,003 [DEBUG] No services specified
2016-05-27 04:01:56,089 [INFO] Running config getCookbooks
2016-05-27 04:01:56,089 [DEBUG] No packages specified
2016-05-27 04:01:56,089 [DEBUG] No groups specified
2016-05-27 04:01:56,089 [DEBUG] No users specified
2016-05-27 04:01:56,090 [DEBUG] No sources specified
2016-05-27 04:01:56,090 [DEBUG] No files specified
2016-05-27 04:01:56,090 [DEBUG] Running command berk
2016-05-27 04:01:56,090 [DEBUG] No test for command berk
2016-05-27 04:02:23,989 [INFO] Command berk succeeded
2016-05-27 04:02:23,989 [DEBUG] Command berk output: Resolving cookbook dependencies...
Fetching 'cfncluster' from source at .
Fetching cookbook index from https://supermarket.getchef.com...
Installing 7-zip (1.0.2) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing apt (3.0.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing build-essential (2.3.1) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Using cfncluster (1.2.2) from source at .
Installing chef_handler (1.4.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing hostname (0.3.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing hostsfile (2.4.5) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing iptables (1.1.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing line (0.6.3) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing nfs (2.2.8) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing ohai (3.0.1) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing openssh (1.6.1) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing poise (2.7.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing poise-archive (1.2.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing poise-languages (1.4.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing poise-python (1.2.1) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing selinux (0.9.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing sysctl (0.7.5) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing tar (0.7.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing windows (1.41.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing yum (3.10.0) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Installing yum-epel (0.6.6) from https://supermarket.getchef.com/ ([opscode] https://supermarket.chef.io:443/api/v1)
Vendoring 7-zip (1.0.2) to /etc/chef/cookbooks/7-zip
Vendoring apt (3.0.0) to /etc/chef/cookbooks/apt
Vendoring build-essential (2.3.1) to /etc/chef/cookbooks/build-essential
Vendoring cfncluster (1.2.2) to /etc/chef/cookbooks/cfncluster
Vendoring chef_handler (1.4.0) to /etc/chef/cookbooks/chef_handler
Vendoring hostname (0.3.0) to /etc/chef/cookbooks/hostname
Vendoring hostsfile (2.4.5) to /etc/chef/cookbooks/hostsfile
Vendoring iptables (1.1.0) to /etc/chef/cookbooks/iptables
Vendoring line (0.6.3) to /etc/chef/cookbooks/line
Vendoring nfs (2.2.8) to /etc/chef/cookbooks/nfs
Vendoring ohai (3.0.1) to /etc/chef/cookbooks/ohai
Vendoring openssh (1.6.1) to /etc/chef/cookbooks/openssh
Vendoring poise (2.7.0) to /etc/chef/cookbooks/poise
Vendoring poise-archive (1.2.0) to /etc/chef/cookbooks/poise-archive
Vendoring poise-languages (1.4.0) to /etc/chef/cookbooks/poise-languages
Vendoring poise-python (1.2.1) to /etc/chef/cookbooks/poise-python
Vendoring selinux (0.9.0) to /etc/chef/cookbooks/selinux
Vendoring sysctl (0.7.5) to /etc/chef/cookbooks/sysctl
Vendoring tar (0.7.0) to /etc/chef/cookbooks/tar
Vendoring windows (1.41.0) to /etc/chef/cookbooks/windows
Vendoring yum (3.10.0) to /etc/chef/cookbooks/yum
Vendoring yum-epel (0.6.6) to /etc/chef/cookbooks/yum-epel
/bin/sh: 1: cd: can't cd to /tmp/cookbooks/cookbook.tar.gz
Resolving cookbook dependencies...
Fetching 'cfncluster' from source at .
Using 7-zip (1.0.2)
Using apt (3.0.0)
Using build-essential (2.3.1)
Using chef_handler (1.4.0)
Using cfncluster (1.2.2) from source at .
Using hostname (0.3.0)
Using hostsfile (2.4.5)
Using iptables (1.1.0)
Using line (0.6.3)
Using nfs (2.2.8)
Using ohai (3.0.1)
Using openssh (1.6.1)
Using poise (2.7.0)
Using poise-archive (1.2.0)
Using poise-languages (1.4.0)
Using poise-python (1.2.1)
Using selinux (0.9.0)
Using sysctl (0.7.5)
Using tar (0.7.0)
Using windows (1.41.0)
Using yum (3.10.0)
Using yum-epel (0.6.6)
Vendoring 7-zip (1.0.2) to /etc/chef/cookbooks/7-zip
Vendoring apt (3.0.0) to /etc/chef/cookbooks/apt
Vendoring build-essential (2.3.1) to /etc/chef/cookbooks/build-essential
Vendoring cfncluster (1.2.2) to /etc/chef/cookbooks/cfncluster
Vendoring chef_handler (1.4.0) to /etc/chef/cookbooks/chef_handler
Vendoring hostname (0.3.0) to /etc/chef/cookbooks/hostname
Vendoring hostsfile (2.4.5) to /etc/chef/cookbooks/hostsfile
Vendoring iptables (1.1.0) to /etc/chef/cookbooks/iptables
Vendoring line (0.6.3) to /etc/chef/cookbooks/line
Vendoring nfs (2.2.8) to /etc/chef/cookbooks/nfs
Vendoring ohai (3.0.1) to /etc/chef/cookbooks/ohai
Vendoring openssh (1.6.1) to /etc/chef/cookbooks/openssh
Vendoring poise (2.7.0) to /etc/chef/cookbooks/poise
Vendoring poise-archive (1.2.0) to /etc/chef/cookbooks/poise-archive
Vendoring poise-languages (1.4.0) to /etc/chef/cookbooks/poise-languages
Vendoring poise-python (1.2.1) to /etc/chef/cookbooks/poise-python
Vendoring selinux (0.9.0) to /etc/chef/cookbooks/selinux
Vendoring sysctl (0.7.5) to /etc/chef/cookbooks/sysctl
Vendoring tar (0.7.0) to /etc/chef/cookbooks/tar
Vendoring windows (1.41.0) to /etc/chef/cookbooks/windows
Vendoring yum (3.10.0) to /etc/chef/cookbooks/yum
Vendoring yum-epel (0.6.6) to /etc/chef/cookbooks/yum-epel
2016-05-27 04:02:23,989 [DEBUG] No services specified
2016-05-27 04:02:24,054 [INFO] Running config chefPrepEnv
2016-05-27 04:02:24,055 [DEBUG] No packages specified
2016-05-27 04:02:24,055 [DEBUG] No groups specified
2016-05-27 04:02:24,055 [DEBUG] No users specified
2016-05-27 04:02:24,055 [DEBUG] No sources specified
2016-05-27 04:02:24,055 [DEBUG] No files specified
2016-05-27 04:02:24,055 [DEBUG] Running command chef
2016-05-27 04:02:24,055 [DEBUG] No test for command chef
2016-05-27 04:02:37,087 [INFO] Command chef succeeded
2016-05-27 04:02:37,088 [DEBUG] Command chef output: Starting Chef Client, version 12.8.1
[2016-05-27T04:02:32+00:00] WARN: Run List override has been provided.
[2016-05-27T04:02:32+00:00] WARN: Original Run List: [recipe[cfncluster::torque_config]]
[2016-05-27T04:02:32+00:00] WARN: Overridden Run List: [recipe[cfncluster::_prep_env]]
resolving cookbooks for run list: ["cfncluster::_prep_env"]
Synchronizing Cookbooks:
- cfncluster (1.2.2)
- build-essential (2.3.1)
- poise-python (1.2.1)
- tar (0.7.0)
- selinux (0.9.0)
- nfs (2.2.8)
- yum (3.10.0)
- yum-epel (0.6.6)
- openssh (1.6.1)
- apt (3.0.0)
- hostname (0.3.0)
- 7-zip (1.0.2)
- line (0.6.3)
- poise (2.7.0)
- poise-languages (1.4.0)
- iptables (1.1.0)
- hostsfile (2.4.5)
- sysctl (0.7.5)
- poise-archive (1.2.0)
- windows (1.41.0)
- ohai (3.0.1)
- chef_handler (1.4.0)
Installing Cookbook Gems:
Compiling Cookbooks...
Converging 7 resources
Recipe: cfncluster::_prep_env
* directory[/etc/cfncluster] action create (up to date)
* directory[/opt/cfncluster] action create (up to date)
* directory[/opt/cfncluster/scripts] action create (up to date)
* template[/etc/cfncluster/cfnconfig] action create
- create new file /etc/cfncluster/cfnconfig
- update content in file /etc/cfncluster/cfnconfig from none to eaea7d
--- /etc/cfncluster/cfnconfig 2016-05-27 04:02:36.723904999 +0000
+++ /etc/cfncluster/.chef-cfnconfig20160527-1515-22lej4 2016-05-27 04:02:36.723904999 +0000
@@ -1 +1,17 @@
+stack_name=cfncluster-pepper
+cfn_preinstall=NONE
+cfn_preinstall_args=NONE
+cfn_postinstall=https://raw.githubusercontent.com/guicho271828/torque-utils/master/setup/test.py
+cfn_postinstall_args="NONE"
+cfn_region=us-west-2
+cfn_scheduler=torque
+cfn_scheduler_slots=vcpus
+cfn_instance_slots=1
+cfn_encrypted_ephemeral=false
+cfn_ephemeral_dir=/scratch
+cfn_shared_dir=/shared
+cfn_proxy=NONE
+cfn_node_type=MasterServer
+cfn_cluster_user=ubuntu
+cfn_volume=vol-44afecf6
- change mode from '' to '0644'
* link[/opt/cfncluster/cfnconfig] action create
- create symlink at /opt/cfncluster/cfnconfig to /etc/cfncluster/cfnconfig
* cookbook_file[fetch_and_run] action create
- create new file /opt/cfncluster/scripts/fetch_and_run
- update content in file /opt/cfncluster/scripts/fetch_and_run from none to 961f37
--- /opt/cfncluster/scripts/fetch_and_run 2016-05-27 04:02:36.727907000 +0000
+++ /opt/cfncluster/scripts/.chef-fetch_and_run20160527-1515-3po2hb 2016-05-27 04:02:36.727907000 +0000
@@ -1 +1,71 @@
+#!/bin/bash
+
+. /etc/cfncluster/cfnconfig
+
+# Error exit function
+function error_exit () {
+ script=`basename $0`
+ echo "cfncluster: $script - $1"
+ logger -t cfncluster "$script - $1"
+ exit 1
+}
+
+function run_preinstall () {
+ RC=0
+ if [ "${cfn_preinstall}" != "NONE" ]; then
+ scheme=$(echo "${cfn_preinstall}"| cut -d: -f1)
+ tmpfile=$(mktemp)
+ if [ "${scheme}" == "s3" ]; then
+ aws --region ${cfn_region} s3 cp ${cfn_preinstall} - > $tmpfile || RC=1
+ else
+ wget -qO- ${cfn_preinstall} > $tmpfile || RC=1
+ fi
+ if [ "${cfn_preinstall_args}" != "NONE" ]; then
+ args=${cfn_preinstall_args}
+ fi
+ /bin/sh $tmpfile $args || RC=1
+ /bin/rm $tmpfile
+ fi
+ if [ $RC -ne 0 ]; then
+ error_exit "Failed to run boot_as_master preinstall"
+ fi
+}
+
+function run_postinstall () {
+ RC=0
+ if [ "${cfn_postinstall}" != "NONE" ]; then
+ scheme=$(echo "${cfn_postinstall}"| cut -d: -f1)
+ tmpfile=$(mktemp)
+ if [ "${scheme}" == "s3" ]; then
+ aws --region ${cfn_region} s3 cp ${cfn_postinstall} - > $tmpfile || RC=1
+ else
+ wget -qO- ${cfn_postinstall} > $tmpfile || RC=1
+ fi
+ if [ "${cfn_postinstall_args}" != "NONE" ]; then
+ args=${cfn_postinstall_args}
+ fi
+ /bin/sh $tmpfile $args || RC=1
+ /bin/rm $tmpfile
+ fi
+ if [ $RC -ne 0 ]; then
+ error_exit "Failed to run boot_as_master postinstall"
+ fi
+}
+
+ACTION=${1#?}
+
+case $ACTION in
+ preinstall)
+ run_preinstall
+ ;;
+
+ postinstall)
+ run_postinstall
+ ;;
+
+ *)
+ echo "Unknown action. Exit gracefully"
+ exit 0
+
+esac
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
* cookbook_file[compute_ready] action create
- create new file /opt/cfncluster/scripts/compute_ready
- update content in file /opt/cfncluster/scripts/compute_ready from none to 3273c9
--- /opt/cfncluster/scripts/compute_ready 2016-05-27 04:02:36.735910999 +0000
+++ /opt/cfncluster/scripts/.chef-compute_ready20160527-1515-474mh7 2016-05-27 04:02:36.735910999 +0000
@@ -1 +1,9 @@
+#!/bin/bash
+
+. /etc/cfncluster/cfnconfig
+
+# Notify compute is ready
+instance_id_url="http://169.254.169.254/latest/meta-data/instance-id"
+instance_id=$(curl --retry 3 --retry-delay 0 --silent --fail ${instance_id_url})
+aws --region ${cfn_region} sqs send-message --queue-url ${cfn_sqs_queue} --message-body '{"Type" : "Notification", "Message" : "{\"StatusCode\":\"Complete\",\"Description\":\"Succesfully launched '${instance_id}'\",\"Event\":\"cfncluster:COMPUTE_READY\",\"EC2InstanceId\":\"'${instance_id}'\",\"Slots\":\"'${cfn_instance_slots}'\"}"}'
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
[2016-05-27T04:02:36+00:00] WARN: Skipping final node save because override_runlist was given
Running handlers:
Running handlers complete
Chef Client finished, 4/7 resources updated in 07 seconds
2016-05-27 04:02:37,088 [DEBUG] No services specified
2016-05-27 04:02:37,315 [INFO] Running config shellRunPreInstall
2016-05-27 04:02:37,315 [DEBUG] No packages specified
2016-05-27 04:02:37,315 [DEBUG] No groups specified
2016-05-27 04:02:37,315 [DEBUG] No users specified
2016-05-27 04:02:37,315 [DEBUG] No sources specified
2016-05-27 04:02:37,316 [DEBUG] No files specified
2016-05-27 04:02:37,316 [DEBUG] Running command runpreinstall
2016-05-27 04:02:37,316 [DEBUG] No test for command runpreinstall
2016-05-27 04:02:37,319 [INFO] Command runpreinstall succeeded
2016-05-27 04:02:37,319 [DEBUG] Command runpreinstall output:
2016-05-27 04:02:37,320 [DEBUG] No services specified
2016-05-27 04:02:37,321 [INFO] Running config chefConfig
2016-05-27 04:02:37,322 [DEBUG] No packages specified
2016-05-27 04:02:37,322 [DEBUG] No groups specified
2016-05-27 04:02:37,322 [DEBUG] No users specified
2016-05-27 04:02:37,322 [DEBUG] No sources specified
2016-05-27 04:02:37,322 [DEBUG] No files specified
2016-05-27 04:02:37,322 [DEBUG] Running command chef
2016-05-27 04:02:37,322 [DEBUG] No test for command chef
2016-05-27 04:04:52,360 [INFO] Command chef succeeded
2016-05-27 04:04:52,360 [DEBUG] Command chef output: Starting Chef Client, version 12.8.1
resolving cookbooks for run list: ["cfncluster::torque_config"]
Synchronizing Cookbooks:
- tar (0.7.0)
- yum-epel (0.6.6)
- cfncluster (1.2.2)
- build-essential (2.3.1)
- hostname (0.3.0)
- yum (3.10.0)
- 7-zip (1.0.2)
- nfs (2.2.8)
- selinux (0.9.0)
- poise-python (1.2.1)
- openssh (1.6.1)
- iptables (1.1.0)
- line (0.6.3)
- apt (3.0.0)
- poise (2.7.0)
- sysctl (0.7.5)
- hostsfile (2.4.5)
- poise-languages (1.4.0)
- windows (1.41.0)
- poise-archive (1.2.0)
- ohai (3.0.1)
- chef_handler (1.4.0)
Installing Cookbook Gems:
Compiling Cookbooks...
[2016-05-27T04:02:42+00:00] WARN: Cloning resource attributes for directory[/shared] from prior resource (CHEF-3694)
[2016-05-27T04:02:42+00:00] WARN: Previous directory[/shared]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/_master_base_config.rb:54:in `from_file'
[2016-05-27T04:02:42+00:00] WARN: Current directory[/shared]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/_master_base_config.rb:72:in `from_file'
[2016-05-27T04:02:42+00:00] WARN: Cloning resource attributes for bash[make install] from prior resource (CHEF-3694)
[2016-05-27T04:02:42+00:00] WARN: Previous bash[make install]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/munge_install.rb:33:in `from_file'
[2016-05-27T04:02:42+00:00] WARN: Current bash[make install]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/torque_install.rb:35:in `from_file'
[2016-05-27T04:02:42+00:00] WARN: Cloning resource attributes for service[munge] from prior resource (CHEF-3694)
[2016-05-27T04:02:42+00:00] WARN: Previous service[munge]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/torque_config.rb:62:in `from_file'
[2016-05-27T04:02:42+00:00] WARN: Current service[munge]: /etc/chef/local-mode-cache/cache/cookbooks/cfncluster/recipes/_master_torque_config.rb:35:in `from_file'
Converging 159 resources
Recipe: apt::default
* file[/var/lib/apt/periodic/update-success-stamp] action nothing (skipped due to action :nothing)
* directory[/etc/apt/apt.conf.d] action create (up to date)
* cookbook_file[/etc/apt/apt.conf.d/15update-stamp] action create (up to date)
* execute[apt-get update] action nothing (skipped due to action :nothing)
* execute[apt-get autoremove] action nothing (skipped due to action :nothing)
* execute[apt-get autoclean] action nothing (skipped due to action :nothing)
* execute[apt-get-update-periodic] action run
- execute apt-get update
* file[/var/lib/apt/periodic/update-success-stamp] action touch
- update utime on file /var/lib/apt/periodic/update-success-stamp
* directory[/var/cache/local] action create (up to date)
* directory[/var/cache/local/preseeding] action create (up to date)
* template[/etc/apt/apt.conf.d/10recommends] action create (up to date)
* apt_package[apt-transport-https] action install (up to date)
Recipe: build-essential::_debian
* apt_package[autoconf] action install (up to date)
* apt_package[binutils-doc] action install (up to date)
* apt_package[bison] action install (up to date)
* apt_package[build-essential] action install (up to date)
* apt_package[flex] action install (up to date)
* apt_package[gettext] action install (up to date)
* apt_package[ncurses-dev] action install (up to date)
Recipe: cfncluster::_setup_python
* python_runtime[2] action install
* poise_languages_system[python2.7] action install
(up to date)
(up to date)
* python_runtime_pip[2] action install (up to date)
* python_package[setuptools] action install (up to date)
* python_package[wheel] action install (up to date)
* python_package[virtualenv] action install (up to date)
(up to date)
Recipe: openssh::default
* apt_package[openssh-client] action install (up to date)
* apt_package[openssh-server] action install (up to date)
* template[/etc/ssh/ssh_config] action create
- update content in file /etc/ssh/ssh_config from 773e13 to 16710c
--- /etc/ssh/ssh_config 2016-03-23 23:48:53.167414403 +0000
+++ /etc/ssh/.chef-ssh_config20160527-1708-15vvoyl 2016-05-27 04:03:34.220638999 +0000
@@ -1,4 +1,4 @@
-# This file was generated by Chef for ip-192-168-0-68.us-west-2.compute.internal
+# This file was generated by Chef for ip-172-31-29-70.us-west-2.compute.internal
# Do NOT modify this file by hand!
Host *
* template[/etc/ssh/sshd_config] action create
- update content in file /etc/ssh/sshd_config from 60c93f to 9af5d3
--- /etc/ssh/sshd_config 2016-03-23 23:48:53.179414403 +0000
+++ /etc/ssh/.chef-sshd_config20160527-1708-9gfp6q 2016-05-27 04:03:34.228643000 +0000
@@ -1,4 +1,4 @@
-# This file was generated by Chef for ip-192-168-0-68.us-west-2.compute.internal
+# This file was generated by Chef for ip-172-31-29-70.us-west-2.compute.internal
# Do NOT modify this file by hand!
ChallengeResponseAuthentication no
* execute[sshd-config-check] action run
- execute /usr/sbin/sshd -t
* execute[sshd-config-check] action nothing (skipped due to action :nothing)
* service[ssh] action enable (up to date)
* service[ssh] action start (up to date)
Recipe: cfncluster::base_install
* selinux_state[SELinux Disabled] action disabled (skipped due to only_if)
* directory[/etc/cfncluster] action create (up to date)
* directory[/opt/cfncluster] action create (up to date)
* directory[/opt/cfncluster/sources] action create (up to date)
* directory[/opt/cfncluster/scripts] action create (up to date)
* python_package[awscli] action install (up to date)
Recipe: nfs::_common
* apt_package[nfs-common] action install (up to date)
* apt_package[rpcbind] action install (up to date)
* directory[/etc/default] action create (skipped due to only_if)
* template[/etc/default/nfs-common] action create
- update content in file /etc/default/nfs-common from b69b0c to c3ba00
--- /etc/default/nfs-common 2016-03-23 23:49:02.819414403 +0000
+++ /etc/default/.chef-nfs-common20160527-1708-sdo009 2016-05-27 04:03:35.325190999 +0000
@@ -1,3 +1,3 @@
-# Generated by Chef for ip-192-168-0-68.us-west-2.compute.internal# Local modifications will be overwritten.
+# Generated by Chef for ip-172-31-29-70.us-west-2.compute.internal# Local modifications will be overwritten.
STATDOPTS="--port 32765 --outgoing-port 32766"
* service[portmap] action restart
- restart service service[portmap]
* service[lock] action restart
- restart service service[lock]
* template[/etc/modprobe.d/lockd.conf] action create
- update content in file /etc/modprobe.d/lockd.conf from d3bb2d to c8330d
--- /etc/modprobe.d/lockd.conf 2016-03-23 23:49:02.855414403 +0000
+++ /etc/modprobe.d/.chef-lockd.conf20160527-1708-1t9ljq1 2016-05-27 04:03:35.385220999 +0000
@@ -1,4 +1,4 @@
-# Generated by Chef for ip-192-168-0-68.us-west-2.compute.internal
+# Generated by Chef for ip-172-31-29-70.us-west-2.compute.internal
# Local modifications will be overwritten.
options lockd nlm_udpport=32768 nlm_tcpport=32768
* service[portmap] action restart
- restart service service[portmap]
* service[lock] action restart
- restart service service[lock]
* service[portmap] action start (up to date)
* service[portmap] action enable (up to date)
* service[lock] action start (up to date)
* service[lock] action enable (up to date)
Recipe: cfncluster::base_install
* service[rpcbind] action start (skipped due to only_if)
* service[rpcbind] action enable (skipped due to only_if)
Recipe: nfs::server
* apt_package[nfs-kernel-server] action install (up to date)
* template[/etc/default/nfs-kernel-server] action create
- update content in file /etc/default/nfs-kernel-server from 1f7ea7 to c63b84
--- /etc/default/nfs-kernel-server 2016-03-23 23:49:04.447414403 +0000
+++ /etc/default/.chef-nfs-kernel-server20160527-1708-1xp8vtr 2016-05-27 04:03:35.497277000 +0000
@@ -1,4 +1,4 @@
-# Generated by Chef for ip-192-168-0-68.us-west-2.compute.internal# Local modifications will be overwritten.
+# Generated by Chef for ip-172-31-29-70.us-west-2.compute.internal# Local modifications will be overwritten.
# Rendered Debian/Ubuntu template variant
RPCMOUNTDOPTS="-p 32767"
RPCNFSDCOUNT="8"
* service[nfs-kernel-server] action start
- start service service[nfs-kernel-server]
* service[nfs-kernel-server] action enable (up to date)
Recipe: nfs::_idmap
* template[/etc/idmapd.conf] action create (up to date)
* service[idmap] action start (up to date)
* service[idmap] action enable (up to date)
Recipe: cfncluster::base_install
* cookbook_file[configure-pat.sh] action create (up to date)
* cookbook_file[setup-ephemeral-drives.sh] action create (up to date)
Recipe: cfncluster::_ec2_udev_rules
* cookbook_file[ec2-volid.rules] action create (up to date)
* cookbook_file[ec2_dev_2_volid.py] action create (up to date)
* cookbook_file[ec2blkdev-init] action create (up to date)
* cookbook_file[attachVolume.py] action create (up to date)
* service[ec2blkdev] action enable (up to date)
* service[ec2blkdev] action start
- start service service[ec2blkdev]
Recipe: cfncluster::base_install
* remote_file[/usr/bin/ec2-metadata] action create (up to date)
* python_package[cfncluster-node] action install
- install version 1.1.3 of package cfncluster-node
* python_package[supervisor] action install (up to date)
* cookbook_file[supervisord.conf] action create (up to date)
* cookbook_file[supervisord-init] action create (up to date)
* cookbook_file[jq-1.4] action create (up to date)
* apt_package[vim] action install (up to date)
* apt_package[ksh] action install (up to date)
* apt_package[tcsh] action install (up to date)
* apt_package[zsh] action install (up to date)
* apt_package[libssl-dev] action install (up to date)
* apt_package[ncurses-dev] action install (up to date)
* apt_package[libpam-dev] action install (up to date)
* apt_package[net-tools] action install (up to date)
* apt_package[libXmu-dev] action install (up to date)
* apt_package[libhwloc-dev] action install (up to date)
* apt_package[tcl-dev] action install (up to date)
* apt_package[automake] action install (up to date)
* apt_package[autoconf] action install (up to date)
* apt_package[python-parted] action install (up to date)
* apt_package[libtool] action install (up to date)
* apt_package[librrd-dev] action install (up to date)
* apt_package[libapr1-dev] action install (up to date)
* apt_package[libconfuse-dev] action install (up to date)
* apt_package[apache2] action install (up to date)
* apt_package[libboost-dev] action install (up to date)
* apt_package[libdb-dev] action install (up to date)
* apt_package[tcsh] action install (up to date)
* apt_package[libssl-dev] action install (up to date)
* apt_package[libncurses5-dev] action install (up to date)
* apt_package[libpam0g-dev] action install (up to date)
* apt_package[libxt-dev] action install (up to date)
* apt_package[libmotif-dev] action install (up to date)
* apt_package[libxmu-dev] action install (up to date)
* apt_package[libxft-dev] action install (up to date)
* apt_package[libhwloc-dev] action install (up to date)
* apt_package[man-db] action install (up to date)
* apt_package[lvm2] action install (up to date)
* apt_package[libmpich-dev] action install (up to date)
* apt_package[libopenmpi-dev] action install (up to date)
* apt_package[r-base] action install (up to date)
* apt_package[libatlas-dev] action install (up to date)
* apt_package[liblas-dev] action install (up to date)
* apt_package[libfftw3-dev] action install (up to date)
Recipe: cfncluster::_ganglia_install
* apt_package[ganglia-monitor] action install (up to date)
* apt_package[rrdtool] action install (up to date)
* apt_package[gmetad] action install (up to date)
* apt_package[ganglia-webfrontend] action install (up to date)
* execute[copy ganglia apache conf] action run (skipped due to not_if)
Recipe: hostname::default
* file[/etc/hostname] action create (up to date)
* execute[hostname ip-172-31-29-70] action run (skipped due to only_if)
* hostsfile_entry[localhost] action append
Recipe: <Dynamically Defined Resource>
* file[/etc/hosts] action create
- update content in file /etc/hosts from aa4ea9 to 5b4b04
--- /etc/hosts 2016-03-14 20:39:48.000000000 +0000
+++ /etc/.chef-hosts20160527-1708-reh3vv 2016-05-27 04:04:00.369707000 +0000
@@ -1,10 +1,17 @@
-127.0.0.1 localhost
+#
+# This file is managed by Chef, using the hostsfile cookbook.
+# Editing this file by hand is highly discouraged!
+#
+# Comments containing an @ sign should not be modified or else
+# hostsfile will be unable to guarantee relative priority in
+# future Chef runs!
+#
-# The following lines are desirable for IPv6 capable hosts
-::1 ip6-localhost ip6-loopback
-fe00::0 ip6-localnet
-ff00::0 ip6-mcastprefix
-ff02::1 ip6-allnodes
-ff02::2 ip6-allrouters
-ff02::3 ip6-allhosts
+127.0.0.1 localhost
+ff02::3 ip6-allhosts
+ff02::1 ip6-allnodes
+ff02::2 ip6-allrouters
+::1 ip6-localhost ip6-loopback
+fe00:: ip6-localnet
+ff00:: ip6-mcastprefix
- Append hostsfile_entry[localhost]
Recipe: hostname::default
* hostsfile_entry[set hostname] action create
Recipe: <Dynamically Defined Resource>
* file[/etc/hosts] action create
- update content in file /etc/hosts from 5b4b04 to 865b0e
--- /etc/hosts 2016-05-27 04:04:00.369707000 +0000
+++ /etc/.chef-hosts20160527-1708-jedai5 2016-05-27 04:04:00.385715000 +0000
@@ -8,6 +8,7 @@
#
127.0.0.1 localhost
+172.31.29.70 ip-172-31-29-70.us-west-2.compute.internal ip-172-31-29-70
ff02::3 ip6-allhosts
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters
- Create hostsfile_entry[set hostname]
Recipe: hostname::default
* ohai[reload] action reload
- re-run ohai and merge results into node attributes
* ohai[reload] action nothing (skipped due to action :nothing)
Recipe: cfncluster::base_config
* execute[setup ephemeral] action run
- execute /usr/local/sbin/setup-ephemeral-drives.sh
Recipe: cfncluster::_master_base_config
* execute[run_configure-pat] action run
- execute /usr/local/sbin/configure-pat.sh
* execute[add_configure-pat] action run
- execute echo -e "\n# Enable PAT\n/usr/local/sbin/configure-pat.sh\n\n" >> /etc/rc.local
* execute[attach_volume] action run
- execute /usr/local/sbin/attachVolume.py vol-44afecf6
* ruby_block[sleeping_for_volume] action run
- execute the ruby block sleeping_for_volume
* ruby_block[setup_disk] action run
- execute the ruby block setup_disk
* ruby_block[sleeping_for_volume] action run
- execute the ruby block sleeping_for_volume
* ruby_block[setup_disk] action run
- execute the ruby block setup_disk
* ruby_block[setup_disk] action run
- execute the ruby block setup_disk
* directory[/shared] action create
- create new directory /shared
- change mode from '' to '01777'
- change owner from '' to 'root'
- change group from '' to 'root'
* mount[/shared] action mount
- mount /dev/disk/by-ebs-volumeid/vol-44afecf6 to /shared
* mount[/shared] action enable
- enable /dev/disk/by-ebs-volumeid/vol-44afecf6
* directory[/shared] action create
- change mode from '0755' to '01777'
* nfs_export[/shared] action create
* execute[exportfs] action nothing (skipped due to action :nothing)
* append_if_no_line[export /shared] action edit
* execute[exportfs] action run
- execute exportfs -ar
* nfs_export[/home] action create
* execute[exportfs] action nothing (skipped due to action :nothing)
* append_if_no_line[export /home] action edit
* execute[exportfs] action run
- execute exportfs -ar
* template[/etc/ganglia/gmetad.conf] action create
- update content in file /etc/ganglia/gmetad.conf from b8f766 to 239275
--- /etc/ganglia/gmetad.conf 2013-10-15 11:51:49.000000000 +0000
+++ /etc/ganglia/.chef-gmetad.conf20160527-1708-13xpkkf 2016-05-27 04:04:08.978009000 +0000
@@ -41,7 +41,7 @@
# data_source "my grid" 50 1.3.4.7:8655 grid.org:8651 grid-backup.org:8651
# data_source "another source" 1.3.4.7:8655 1.3.4.8
-data_source "my cluster" localhost
+data_source "cfncluster-pepper" localhost
#
# Round-Robin Archives
@@ -69,7 +69,7 @@
# The name of this Grid. All the data sources above will be wrapped in a GRID
# tag with this name.
# default: unspecified
-# gridname "MyGrid"
+#gridname ""
#
#-------------------------------------------------------------------------------
# The authority URL for this grid. Used by other gmetads to locate graphs
@@ -150,39 +150,18 @@
# default: unspecified
# carbon_server "my.graphite.box"
#
-# The port and protocol on which Graphite is listening
+# The port on which Graphite is listening
# default: 2003
# carbon_port 2003
#
-# default: tcp
-# carbon_protocol udp
-#
-# **Deprecated in favor of graphite_path** A prefix to prepend to the
-# metric names exported by gmetad. Graphite uses dot-
+# A prefix to prepend to the metric names exported by gmetad. Graphite uses dot-
# separated paths to organize and refer to metrics.
# default: unspecified
# graphite_prefix "datacenter1.gmetad"
#
-# A user-definable graphite path. Graphite uses dot-
-# separated paths to organize and refer to metrics.
-# For reverse compatibility graphite_prefix will be prepended to this
-# path, but this behavior should be considered deprecated.
-# This path may include 3 variables that will be replaced accordingly:
-# %s -> source (cluster name)
-# %h -> host (host name)
-# %m -> metric (metric name)
-# default: graphite_prefix.%s.%h.%m
-# graphite_path "datacenter1.gmetad.%s.%h.%m
-
# Number of milliseconds gmetad will wait for a response from the graphite server
# default: 500
# carbon_timeout 500
-
-#-------------------------------------------------------------------------------
-# Memcached configuration (if it has been compiled in)
-# Format documentation at http://docs.libmemcached.org/libmemcached_configuration.html
-# default: ""
-# memcached_parameters "--SERVER=127.0.0.1"
#
* template[/etc/ganglia/gmond.conf] action create
- update content in file /etc/ganglia/gmond.conf from 556740 to 5278eb
--- /etc/ganglia/gmond.conf 2013-10-15 11:51:49.000000000 +0000
+++ /etc/ganglia/.chef-gmond.conf20160527-1708-1wpph5g 2016-05-27 04:04:08.986012999 +0000
@@ -1,338 +1,361 @@
-/* This configuration is as close to 2.5.x default behavior as possible
- The values closely match ./gmond/metric.h definitions in 2.5.x */
-globals {
- daemonize = yes
- setuid = yes
- user = ganglia
- debug_level = 0
- max_udp_msg_len = 1472
- mute = no
- deaf = no
- host_dmax = 0 /*secs */
- cleanup_threshold = 300 /*secs */
- gexec = no
- send_metadata_interval = 0
-}
+/* This configuration is as close to 2.5.x default behavior as possible
+ The values closely match ./gmond/metric.h definitions in 2.5.x */
+globals {
+ daemonize = yes
+ setuid = yes
+ user = nobody
+ debug_level = 0
+ max_udp_msg_len = 1472
+ mute = no
+ deaf = no
+ allow_extra_data = yes
+ host_dmax = 3600 /*secs. Expires (removes from web interface) hosts in 1 hour */
+ host_tmax = 20 /*secs */
+ cleanup_threshold = 300 /*secs */
+ gexec = no
+ send_metadata_interval = 0 /*secs */
+}
-/* If a cluster attribute is specified, then all gmond hosts are wrapped inside
- * of a <CLUSTER> tag. If you do not specify a cluster tag, then all <HOSTS> will
- * NOT be wrapped inside of a <CLUSTER> tag. */
-cluster {
- name = "unspecified"
- owner = "unspecified"
- latlong = "unspecified"
- url = "unspecified"
-}
+/*
+ * The cluster attributes specified will be used as part of the <CLUSTER>
+ * tag that will wrap all hosts collected by this instance.
+ */
+cluster {
+ name = "cfncluster-pepper"
+}
-/* The host section describes attributes of the host, like the location */
-host {
- location = "unspecified"
-}
+/* The host section describes attributes of the host, like the location */
+host {
+ location = "us-west-2b"
+}
-/* Feel free to specify as many udp_send_channels as you like. Gmond
- used to only support having a single channel */
-udp_send_channel {
- mcast_join = 239.2.11.71
- port = 8649
- ttl = 1
-}
+/* Feel free to specify as many udp_send_channels as you like. Gmond
+ used to only support having a single channel */
+udp_send_channel {
+ bind_hostname = yes # Highly recommended, soon to be default.
+ # This option tells gmond to use a source address
+ # that resolves to the machine's hostname. Without
+ # this, the metrics may appear to come from any
+ # interface and the DNS names associated with
+ # those IPs will be used to create the RRDs.
+ host = ip-172-31-29-70
+ port = 8649
+ ttl = 1
+}
-/* You can specify as many udp_recv_channels as you like as well. */
-udp_recv_channel {
- mcast_join = 239.2.11.71
- port = 8649
- bind = 239.2.11.71
-}
+/* You can specify as many udp_recv_channels as you like as well. */
+udp_recv_channel {
+ port = 8649
+}
-/* You can specify as many tcp_accept_channels as you like to share
- an xml description of the state of the cluster */
-tcp_accept_channel {
- port = 8649
-}
+/* You can specify as many tcp_accept_channels as you like to share
+ an xml description of the state of the cluster */
+tcp_accept_channel {
+ port = 8649
+}
-/* Each metrics module that is referenced by gmond must be specified and
- loaded. If the module has been statically linked with gmond, it does not
- require a load path. However all dynamically loadable modules must include
- a load path. */
-modules {
- module {
- name = "core_metrics"
- }
- module {
- name = "cpu_module"
- path = "/usr/lib/ganglia/modcpu.so"
- }
- module {
- name = "disk_module"
- path = "/usr/lib/ganglia/moddisk.so"
- }
- module {
- name = "load_module"
- path = "/usr/lib/ganglia/modload.so"
- }
- module {
- name = "mem_module"
- path = "/usr/lib/ganglia/modmem.so"
- }
- module {
- name = "net_module"
- path = "/usr/lib/ganglia/modnet.so"
- }
- module {
- name = "proc_module"
- path = "/usr/lib/ganglia/modproc.so"
- }
- module {
- name = "sys_module"
- path = "/usr/lib/ganglia/modsys.so"
- }
-}
+/* Channel to receive sFlow datagrams */
+#udp_recv_channel {
+# port = 6343
+#}
-include ('/etc/ganglia/conf.d/*.conf')
+/* Optional sFlow settings */
+#sflow {
+# udp_port = 6343
+# accept_vm_metrics = yes
+# accept_jvm_metrics = yes
+# multiple_jvm_instances = no
+# accept_http_metrics = yes
+# multiple_http_instances = no
+# accept_memcache_metrics = yes
+# multiple_memcache_instances = no
+#}
+/* Each metrics module that is referenced by gmond must be specified and
+ loaded. If the module has been statically linked with gmond, it does
+ not require a load path. However all dynamically loadable modules must
+ include a load path. */
+modules {
+ module {
+ name = "core_metrics"
+ }
+ module {
+ name = "cpu_module"
+ path = "modcpu.so"
+ }
+ module {
+ name = "disk_module"
+ path = "moddisk.so"
+ }
+ module {
+ name = "load_module"
+ path = "modload.so"
+ }
+ module {
+ name = "mem_module"
+ path = "modmem.so"
+ }
+ module {
+ name = "net_module"
+ path = "modnet.so"
+ }
+ module {
+ name = "proc_module"
+ path = "modproc.so"
+ }
+ module {
+ name = "sys_module"
+ path = "modsys.so"
+ }
+}
-/* The old internal 2.5.x metric array has been replaced by the following
- collection_group directives. What follows is the default behavior for
- collecting and sending metrics that is as close to 2.5.x behavior as
+/* The old internal 2.5.x metric array has been replaced by the following
+ collection_group directives. What follows is the default behavior for
+ collecting and sending metrics that is as close to 2.5.x behavior as
possible. */
-/* This collection group will cause a heartbeat (or beacon) to be sent every
- 20 seconds. In the heartbeat is the GMOND_STARTED data which expresses
- the age of the running gmond. */
-collection_group {
- collect_once = yes
- time_threshold = 20
- metric {
- name = "heartbeat"
- }
-}
+/* This collection group will cause a heartbeat (or beacon) to be sent every
+ 20 seconds. In the heartbeat is the GMOND_STARTED data which expresses
+ the age of the running gmond. */
+collection_group {
+ collect_once = yes
+ time_threshold = 20
+ metric {
+ name = "heartbeat"
+ }
+}
-/* This collection group will send general info about this host every 1200 secs.
- This information doesn't change between reboots and is only collected once. */
-collection_group {
- collect_once = yes
- time_threshold = 1200
- metric {
- name = "cpu_num"
- title = "CPU Count"
- }
- metric {
- name = "cpu_speed"
- title = "CPU Speed"
- }
- metric {
- name = "mem_total"
- title = "Memory Total"
- }
- /* Should this be here? Swap can be added/removed between reboots. */
- metric {
- name = "swap_total"
- title = "Swap Space Total"
- }
- metric {
- name = "boottime"
- title = "Last Boot Time"
- }
- metric {
- name = "machine_type"
- title = "Machine Type"
- }
- metric {
- name = "os_name"
- title = "Operating System"
- }
- metric {
- name = "os_release"
- title = "Operating System Release"
- }
- metric {
- name = "location"
- title = "Location"
- }
-}
+/* This collection group will send general info about this host every
+ 1200 secs.
+ This information doesn't change between reboots and is only collected
+ once. */
+collection_group {
+ collect_once = yes
+ time_threshold = 1200
+ metric {
+ name = "cpu_num"
+ title = "CPU Count"
+ }
+ metric {
+ name = "cpu_speed"
+ title = "CPU Speed"
+ }
+ metric {
+ name = "mem_total"
+ title = "Memory Total"
+ }
+ /* Should this be here? Swap can be added/removed between reboots. */
+ metric {
+ name = "swap_total"
+ title = "Swap Space Total"
+ }
+ metric {
+ name = "boottime"
+ title = "Last Boot Time"
+ }
+ metric {
+ name = "machine_type"
+ title = "Machine Type"
+ }
+ metric {
+ name = "os_name"
+ title = "Operating System"
+ }
+ metric {
+ name = "os_release"
+ title = "Operating System Release"
+ }
+ metric {
+ name = "location"
+ title = "Location"
+ }
+}
-/* This collection group will send the status of gexecd for this host every 300 secs */
-/* Unlike 2.5.x the default behavior is to report gexecd OFF. */
-collection_group {
- collect_once = yes
- time_threshold = 300
- metric {
- name = "gexec"
- title = "Gexec Status"
- }
-}
+/* This collection group will send the status of gexecd for this host
+ every 300 secs.*/
+/* Unlike 2.5.x the default behavior is to report gexecd OFF. */
+collection_group {
+ collect_once = yes
+ time_threshold = 300
+ metric {
+ name = "gexec"
+ title = "Gexec Status"
+ }
+}
-/* This collection group will collect the CPU status info every 20 secs.
- The time threshold is set to 90 seconds. In honesty, this time_threshold could be
- set significantly higher to reduce unneccessary network chatter. */
-collection_group {
- collect_every = 20
- time_threshold = 90
- /* CPU status */
- metric {
- name = "cpu_user"
- value_threshold = "1.0"
- title = "CPU User"
- }
- metric {
- name = "cpu_system"
- value_threshold = "1.0"
- title = "CPU System"
- }
- metric {
- name = "cpu_idle"
- value_threshold = "5.0"
- title = "CPU Idle"
- }
- metric {
- name = "cpu_nice"
- value_threshold = "1.0"
- title = "CPU Nice"
- }
- metric {
- name = "cpu_aidle"
- value_threshold = "5.0"
- title = "CPU aidle"
- }
- metric {
- name = "cpu_wio"
- value_threshold = "1.0"
- title = "CPU wio"
- }
- /* The next two metrics are optional if you want more detail...
- ... since they are accounted for in cpu_system.
- metric {
- name = "cpu_intr"
- value_threshold = "1.0"
- title = "CPU intr"
- }
- metric {
- name = "cpu_sintr"
- value_threshold = "1.0"
- title = "CPU sintr"
- }
- */
-}
+/* This collection group will collect the CPU status info every 20 secs.
+ The time threshold is set to 90 seconds. In honesty, this
+ time_threshold could be set significantly higher to reduce
+ unneccessary network chatter. */
+collection_group {
+ collect_every = 20
+ time_threshold = 90
+ /* CPU status */
+ metric {
+ name = "cpu_user"
+ value_threshold = "1.0"
+ title = "CPU User"
+ }
+ metric {
+ name = "cpu_system"
+ value_threshold = "1.0"
+ title = "CPU System"
+ }
+ metric {
+ name = "cpu_idle"
+ value_threshold = "5.0"
+ title = "CPU Idle"
+ }
+ metric {
+ name = "cpu_nice"
+ value_threshold = "1.0"
+ title = "CPU Nice"
+ }
+ metric {
+ name = "cpu_aidle"
+ value_threshold = "5.0"
+ title = "CPU aidle"
+ }
+ metric {
+ name = "cpu_wio"
+ value_threshold = "1.0"
+ title = "CPU wio"
+ }
+ /* The next two metrics are optional if you want more detail...
+ ... since they are accounted for in cpu_system.
+ metric {
+ name = "cpu_intr"
+ value_threshold = "1.0"
+ title = "CPU intr"
+ }
+ metric {
+ name = "cpu_sintr"
+ value_threshold = "1.0"
+ title = "CPU sintr"
+ }
+ */
+}
-collection_group {
- collect_every = 20
- time_threshold = 90
- /* Load Averages */
- metric {
- name = "load_one"
- value_threshold = "1.0"
- title = "One Minute Load Average"
- }
- metric {
- name = "load_five"
- value_threshold = "1.0"
- title = "Five Minute Load Average"
- }
- metric {
- name = "load_fifteen"
- value_threshold = "1.0"
- title = "Fifteen Minute Load Average"
+collection_group {
+ collect_every = 20
+ time_threshold = 90
+ /* Load Averages */
+ metric {
+ name = "load_one"
+ value_threshold = "1.0"
+ title = "One Minute Load Average"
}
-}
+ metric {
+ name = "load_five"
+ value_threshold = "1.0"
+ title = "Five Minute Load Average"
+ }
+ metric {
+ name = "load_fifteen"
+ value_threshold = "1.0"
+ title = "Fifteen Minute Load Average"
+ }
+}
-/* This group collects the number of running and total processes */
-collection_group {
- collect_every = 80
- time_threshold = 950
- metric {
- name = "proc_run"
- value_threshold = "1.0"
- title = "Total Running Processes"
- }
- metric {
- name = "proc_total"
- value_threshold = "1.0"
- title = "Total Processes"
- }
+/* This group collects the number of running and total processes */
+collection_group {
+ collect_every = 80
+ time_threshold = 950
+ metric {
+ name = "proc_run"
+ value_threshold = "1.0"
+ title = "Total Running Processes"
+ }
+ metric {
+ name = "proc_total"
+ value_threshold = "1.0"
+ title = "Total Processes"
+ }
}
-/* This collection group grabs the volatile memory metrics every 40 secs and
- sends them at least every 180 secs. This time_threshold can be increased
- significantly to reduce unneeded network traffic. */
-collection_group {
- collect_every = 40
- time_threshold = 180
- metric {
- name = "mem_free"
- value_threshold = "1024.0"
- title = "Free Memory"
- }
- metric {
- name = "mem_shared"
- value_threshold = "1024.0"
- title = "Shared Memory"
- }
- metric {
- name = "mem_buffers"
- value_threshold = "1024.0"
- title = "Memory Buffers"
- }
- metric {
- name = "mem_cached"
- value_threshold = "1024.0"
- title = "Cached Memory"
- }
- metric {
- name = "swap_free"
- value_threshold = "1024.0"
- title = "Free Swap Space"
- }
-}
+/* This collection group grabs the volatile memory metrics every 40 secs and
+ sends them at least every 180 secs. This time_threshold can be increased
+ significantly to reduce unneeded network traffic. */
+collection_group {
+ collect_every = 40
+ time_threshold = 180
+ metric {
+ name = "mem_free"
+ value_threshold = "1024.0"
+ title = "Free Memory"
+ }
+ metric {
+ name = "mem_shared"
+ value_threshold = "1024.0"
+ title = "Shared Memory"
+ }
+ metric {
+ name = "mem_buffers"
+ value_threshold = "1024.0"
+ title = "Memory Buffers"
+ }
+ metric {
+ name = "mem_cached"
+ value_threshold = "1024.0"
+ title = "Cached Memory"
+ }
+ metric {
+ name = "swap_free"
+ value_threshold = "1024.0"
+ title = "Free Swap Space"
+ }
+}
-collection_group {
- collect_every = 40
- time_threshold = 300
- metric {
- name = "bytes_out"
- value_threshold = 4096
- title = "Bytes Sent"
- }
- metric {
- name = "bytes_in"
- value_threshold = 4096
- title = "Bytes Received"
- }
- metric {
- name = "pkts_in"
- value_threshold = 256
- title = "Packets Received"
- }
- metric {
- name = "pkts_out"
- value_threshold = 256
- title = "Packets Sent"
- }
+collection_group {
+ collect_every = 40
+ time_threshold = 300
+ metric {
+ name = "bytes_out"
+ value_threshold = 4096
+ title = "Bytes Sent"
+ }
+ metric {
+ name = "bytes_in"
+ value_threshold = 4096
+ title = "Bytes Received"
+ }
+ metric {
+ name = "pkts_in"
+ value_threshold = 256
+ title = "Packets Received"
+ }
+ metric {
+ name = "pkts_out"
+ value_threshold = 256
+ title = "Packets Sent"
+ }
}
-/* Different than 2.5.x default since the old config made no sense */
-collection_group {
- collect_every = 1800
- time_threshold = 3600
- metric {
- name = "disk_total"
- value_threshold = 1.0
- title = "Total Disk Space"
- }
+/* Different than 2.5.x default since the old config made no sense */
+collection_group {
+ collect_every = 1800
+ time_threshold = 3600
+ metric {
+ name = "disk_total"
+ value_threshold = 1.0
+ title = "Total Disk Space"
+ }
}
-collection_group {
- collect_every = 40
- time_threshold = 180
- metric {
- name = "disk_free"
- value_threshold = 1.0
- title = "Disk Space Available"
- }
- metric {
- name = "part_max_used"
- value_threshold = 1.0
- title = "Maximum Disk Space Used"
- }
+collection_group {
+ collect_every = 40
+ time_threshold = 180
+ metric {
+ name = "disk_free"
+ value_threshold = 1.0
+ title = "Disk Space Available"
+ }
+ metric {
+ name = "part_max_used"
+ value_threshold = 1.0
+ title = "Maximum Disk Space Used"
+ }
}
+include ("/etc/ganglia/conf.d/*.conf")
* service[gmetad] action enable (up to date)
* service[gmetad] action start (up to date)
* service[ganglia-monitor] action enable (up to date)
* service[ganglia-monitor] action start (up to date)
* service[apache2] action enable (up to date)
* service[apache2] action start (up to date)
* user[ubuntu] action create
- alter user ubuntu
* bash[ssh-keygen] action run
- execute "bash" "/tmp/chef-script20160527-1708-peccoq"
* bash[copy_and_perms] action run
- execute "bash" "/tmp/chef-script20160527-1708-oybao6"
* bash[ssh-keyscan] action run
- execute "bash" "/tmp/chef-script20160527-1708-67x7m8"
* template[/etc/sqswatcher.cfg] action create
- create new file /etc/sqswatcher.cfg
- update content in file /etc/sqswatcher.cfg from none to 8fc0d0
--- /etc/sqswatcher.cfg 2016-05-27 04:04:09.270155000 +0000
+++ /etc/.chef-sqswatcher.cfg20160527-1708-3nl2q1 2016-05-27 04:04:09.270155000 +0000
@@ -1 +1,7 @@
+[sqswatcher]
+region = us-west-2
+sqsqueue = cfncluster-pepper-SQS-1BT94SFWNBLM2
+table_name = cfncluster-pepper-DynamoDBTable-11ZI0I9INH8UI
+scheduler = torque
+cluster_user = ubuntu
- change mode from '' to '0644'
- change owner from '' to 'root'
- change group from '' to 'root'
Recipe: cfncluster::base_config
* template[/etc/sudoers.d/99-cfncluster-user-tty] action create
- create new file /etc/sudoers.d/99-cfncluster-user-tty
- update content in file /etc/sudoers.d/99-cfncluster-user-tty from none to 1c4b99
--- /etc/sudoers.d/99-cfncluster-user-tty 2016-05-27 04:04:09.274156999 +0000
+++ /etc/sudoers.d/.chef-99-cfncluster-user-tty20160527-1708-17ejlhg 2016-05-27 04:04:09.274156999 +0000
@@ -1 +1,2 @@
+Defaults:ubuntu !requiretty
- change mode from '' to '0600'
- change owner from '' to 'root'
- change group from '' to 'root'
* template[/etc/cfncluster/cfncluster_supervisord.conf] action create
- create new file /etc/cfncluster/cfncluster_supervisord.conf
- update content in file /etc/cfncluster/cfncluster_supervisord.conf from none to 7d6ae9
--- /etc/cfncluster/cfncluster_supervisord.conf 2016-05-27 04:04:09.278158999 +0000
+++ /etc/cfncluster/.chef-cfncluster_supervisord.conf20160527-1708-1cjvkej 2016-05-27 04:04:09.278158999 +0000
@@ -1 +1,7 @@
+# Generated by Chef for cfncluster MasterServer# Local modifications could be be overwritten.
+[program:sqswatcher]
+command = /usr/local/bin/sqswatcher
+redirect_stderr = true
+stdout_logfile = /var/log/sqswatcher
+
- change mode from '' to '0644'
- change owner from '' to 'root'
- change group from '' to 'root'
* service[supervisord] action enable
- enable service service[supervisord]
* service[supervisord] action start
- start service service[supervisord]
Recipe: cfncluster::munge_install
* remote_file[/opt/cfncluster/sources/munge-0.5.11.tar.gz] action create (skipped due to not_if)
* bash[make install] action run (skipped due to not_if)
* cookbook_file[munge-init] action create (up to date)
* user[munge] action create (up to date)
* directory[/var/log/munge] action create (up to date)
* directory[/etc/munge] action create (up to date)
* directory[/var/run/munge] action create
- create new directory /var/run/munge
- change owner from '' to 'munge'
Recipe: cfncluster::torque_install
* apt_package[libboost1.54] action install (up to date)
* apt_package[libboost1.54-dev] action install (up to date)
* remote_file[/opt/cfncluster/sources/torque-6.0.0.tar.gz] action create (skipped due to not_if)
* bash[make install] action run (skipped due to not_if)
* directory[/opt/torque/bin/] action create (up to date)
* directory[/var/spool/torque] action create (up to date)
* cookbook_file[torque.setup] action create (up to date)
Recipe: cfncluster::torque_config
* append_if_no_line[add torque libs to ld.so.conf] action edit (up to date)
* execute[run-ldconfig] action run
- execute ldconfig
* template[/var/spool/torque/server_name] action create
- update content in file /var/spool/torque/server_name from 18888d to 865615
--- /var/spool/torque/server_name 2016-03-23 23:57:04.251414403 +0000
+++ /var/spool/torque/.chef-server_name20160527-1708-134uhdl 2016-05-27 04:04:14.752895405 +0000
@@ -1,2 +1,2 @@
-ip-192-168-0-68
+ip-172-31-29-70
* remote_file[install trqauthd service] action create
- create new file /etc/init.d/trqauthd
- update content in file /etc/init.d/trqauthd from none to 656827
--- /etc/init.d/trqauthd 2016-05-27 04:04:14.904971633 +0000
+++ /etc/init.d/.chef-install trqauthd service20160527-1708-10751qf 2016-05-27 04:04:14.904971633 +0000
@@ -1 +1,119 @@
+#! /bin/sh
+### BEGIN INIT INFO
+# Provides: trqauthd
+# Required-Start: $local_fs
+# Should-Start:
+# Required-Stop:
+# Should-Stop:
+# Default-Start: 2 3 5
+# Default-Stop: S 0 1 6
+# Description: Torque is a versatile batch system for SMPs and clusters
+### END INIT INFO
+
+# TO INSTALL DO (as root):
+# cp debian.trqauthd /etc/init.d/trqauthd && update-rc.d trqauthd defaults
+
+DESC="TORQUE auth daemon"
+NAME=trqauthd
+DAEMON=/opt/torque/sbin/$NAME
+PBS_DAEMON=$DAEMON
+PBS_HOME=/var/spool/torque
+PIDFILE=$PBS_HOME/server_priv/trqauthd.lock
+SCRIPTNAME=/etc/init.d/$NAME
+export PBS_DAEMON PIDFILE
+
+# Exit if the package is not installed
+[ -x "$DAEMON" ] || exit 0
+
+# Read configuration variable file if it is present
+[ -r /etc/default/$NAME ] && . /etc/default/$NAME
+
+# Load the VERBOSE setting and other rcS variables
+[ -f /etc/default/rcS ] && . /etc/default/rcS
+
+# Define LSB log_* functions.
+. /lib/lsb/init-functions
+
+do_start()
+{
+ # Return
+ # 0 if daemon has been started
+ # 1 if daemon was already running
+ # 2 if daemon could not be started
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON --test > /dev/null \
+ || return 1
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON -- \
+ $DAEMON_ARGS \
+ || return 2
+}
+
+do_stop()
+{
+ # Return
+ # 0 if daemon has been stopped
+ # 1 if daemon was already stopped
+ # 2 if daemon could not be stopped
+ # other if a failure occurred
+ start-stop-daemon --stop --quiet --retry=TERM/1/KILL/5 --pidfile $PIDFILE --name $NAME
+ RETVAL="$?"
+ [ "$RETVAL" = 2 ] && return 2
+ start-stop-daemon --stop --quiet --oknodo --retry=0/1/KILL/5 --exec $DAEMON
+ [ "$?" = 2 ] && return 2
+ rm -f $PIDFILE
+ return "$RETVAL"
+}
+
+do_reload() {
+ start-stop-daemon --stop --signal HUP --quiet --pidfile $PIDFILE --name $NAME
+ return 0
+}
+
+case "$1" in
+ start)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Starting $DESC" "$NAME"
+ do_start
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ stop)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Stopping $DESC" "$NAME"
+ do_stop
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ reload|force-reload)
+ log_daemon_msg "Reloading $DESC" "$NAME"
+ do_reload
+ log_end_msg $?
+ ;;
+ restart)
+ log_daemon_msg "Restarting $DESC" "$NAME"
+ do_stop
+ sleep 1
+ case "$?" in
+ 0|1)
+ do_start
+ case "$?" in
+ 0) log_end_msg 0 ;;
+ 1) log_end_msg 1 ;; # Old process is still running
+ *) log_end_msg 1 ;; # Failed to start
+ esac
+ ;;
+ *)
+ # Failed to stop
+ log_end_msg 1
+ ;;
+ esac
+ ;;
+ *)
+ echo "Usage: $SCRIPTNAME {start|stop|restart|reload|force-reload}" >&2
+ exit 3
+ ;;
+esac
+
+:
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
* service[trqauthd] action enable
- enable service service[trqauthd]
* service[trqauthd] action start
- start service service[trqauthd]
* template[/etc/munge/munge.key] action create
- create new file /etc/munge/munge.key
- update content in file /etc/munge/munge.key from none to d1cf51
--- /etc/munge/munge.key 2016-05-27 04:04:15.309172999 +0000
+++ /etc/munge/.chef-munge.key20160527-1708-1vnxtv0 2016-05-27 04:04:15.309172999 +0000
@@ -1 +1,2 @@
+YflQEFLjoxsmEK5vQyKklkLKJ#LkjLKDJF@*(#)ajLKQ@hLKN#()FSU(#@KLJH$@HKSASG)*DUJJDksdN
- change mode from '' to '0600'
- change owner from '' to 'munge'
* service[munge] action enable
- enable service service[munge]
* service[munge] action start
- start service service[munge]
* cookbook_file[/etc/profile.d/torque.sh] action create
- create new file /etc/profile.d/torque.sh
- update content in file /etc/profile.d/torque.sh from none to d8e977
--- /etc/profile.d/torque.sh 2016-05-27 04:04:15.897466999 +0000
+++ /etc/profile.d/.chef-torque.sh20160527-1708-51hnjt 2016-05-27 04:04:15.897466999 +0000
@@ -1 +1,5 @@
+PATH=$PATH:/opt/torque/bin:/opt/torque/sbin
+MANPATH=$MANPATH:/opt/torque/share/man
+
+export PATH MANPATH
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
Recipe: cfncluster::_master_torque_config
* bash[run-torque-setup] action run
- execute "bash" "/tmp/chef-script20160527-1708-1hs0pwl"
* remote_file[install pbs_server service] action create
- create new file /etc/init.d/pbs_server
- update content in file /etc/init.d/pbs_server from none to 3beea6
--- /etc/init.d/pbs_server 2016-05-27 04:04:18.554795000 +0000
+++ /etc/init.d/.chef-install pbs_server service20160527-1708-1t6rzvb 2016-05-27 04:04:18.554795000 +0000
@@ -1 +1,119 @@
+#!/bin/sh
+### BEGIN INIT INFO
+# Provides: pbs_server
+# Required-Start: $local_fs
+# Should-Start:
+# Required-Stop:
+# Should-Stop:
+# Default-Start: 2 3 5
+# Default-Stop: S 0 1 6
+# Description: Torque is a versatile batch system for SMPs and clusters
+### END INIT INFO
+
+# TO INSTALL DO (as root):
+# cp debian.pbs_server /etc/init.d/pbs_server && update-rc.d pbs_server defaults
+
+DESC="TORQUE server"
+NAME=pbs_server
+DAEMON=/opt/torque/sbin/$NAME
+PBS_DAEMON=$DAEMON
+PBS_HOME=/var/spool/torque
+PIDFILE=$PBS_HOME/server_priv/server.lock
+SCRIPTNAME=/etc/init.d/$NAME
+export PBS_DAEMON PBS_HOME PIDFILE
+
+# Exit if the package is not installed
+[ -x "$DAEMON" ] || exit 0
+
+# Read configuration variable file if it is present
+[ -r /etc/default/$NAME ] && . /etc/default/$NAME
+
+# Load the VERBOSE setting and other rcS variables
+[ -f /etc/default/rcS ] && . /etc/default/rcS
+
+# Define LSB log_* functions.
+. /lib/lsb/init-functions
+
+do_start()
+{
+ # Return
+ # 0 if daemon has been started
+ # 1 if daemon was already running
+ # 2 if daemon could not be started
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON --test > /dev/null \
+ || return 1
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON -- \
+ $DAEMON_ARGS \
+ || return 2
+}
+
+do_stop()
+{
+ # Return
+ # 0 if daemon has been stopped
+ # 1 if daemon was already stopped
+ # 2 if daemon could not be stopped
+ # other if a failure occurred
+ start-stop-daemon --stop --quiet --retry=TERM/30/KILL/5 --pidfile $PIDFILE --name $NAME
+ RETVAL="$?"
+ [ "$RETVAL" = 2 ] && return 2
+ start-stop-daemon --stop --quiet --oknodo --retry=0/30/KILL/5 --exec $DAEMON
+ [ "$?" = 2 ] && return 2
+ rm -f $PIDFILE
+ return "$RETVAL"
+}
+
+do_reload() {
+ start-stop-daemon --stop --signal HUP --quiet --pidfile $PIDFILE --name $NAME
+ return 0
+}
+
+case "$1" in
+ start)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Starting $DESC" "$NAME"
+ do_start
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ stop)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Stopping $DESC" "$NAME"
+ do_stop
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ reload|force-reload)
+ log_daemon_msg "Reloading $DESC" "$NAME"
+ do_reload
+ log_end_msg $?
+ ;;
+ restart)
+ log_daemon_msg "Restarting $DESC" "$NAME"
+ do_stop
+ sleep 1
+ case "$?" in
+ 0|1)
+ do_start
+ case "$?" in
+ 0) log_end_msg 0 ;;
+ 1) log_end_msg 1 ;; # Old process is still running
+ *) log_end_msg 1 ;; # Failed to start
+ esac
+ ;;
+ *)
+ # Failed to stop
+ log_end_msg 1
+ ;;
+ esac
+ ;;
+ *)
+ echo "Usage: $SCRIPTNAME {start|stop|restart|reload|force-reload}" >&2
+ exit 3
+ ;;
+esac
+
+:
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
* service[munge] action enable (up to date)
* service[munge] action start (up to date)
* service[pbs_server] action enable
- enable service service[pbs_server]
* service[pbs_server] action restart
- restart service service[pbs_server]
* remote_file[install pbs_sched service] action create
- create new file /etc/init.d/pbs_sched
- update content in file /etc/init.d/pbs_sched from none to 12f9b2
--- /etc/init.d/pbs_sched 2016-05-27 04:04:49.902461001 +0000
+++ /etc/init.d/.chef-install pbs_sched service20160527-1708-17eyvn5 2016-05-27 04:04:49.902461001 +0000
@@ -1 +1,119 @@
+#! /bin/sh
+### BEGIN INIT INFO
+# Provides: pbs_sched
+# Required-Start: $local_fs
+# Should-Start: pbs_server
+# Required-Stop:
+# Should-Stop:
+# Default-Start: 2 3 5
+# Default-Stop: S 0 1 6
+# Description: Torque is a versatile batch system for SMPs and clusters
+### END INIT INFO
+
+# TO INSTALL DO (as root):
+# cp debian.pbs_sched /etc/init.d/pbs_sched && update-rc.d pbs_sched defaults
+
+DESC="TORQUE scheduler"
+NAME=pbs_sched
+DAEMON=/opt/torque/sbin/$NAME
+PBS_DAEMON=$DAEMON
+PBS_HOME=/var/spool/torque
+PIDFILE=$PBS_HOME/sched_priv/sched.lock
+SCRIPTNAME=/etc/init.d/$NAME
+export PBS_DAEMON PIDFILE
+
+# Exit if the package is not installed
+[ -x "$DAEMON" ] || exit 0
+
+# Read configuration variable file if it is present
+[ -r /etc/default/$NAME ] && . /etc/default/$NAME
+
+# Load the VERBOSE setting and other rcS variables
+[ -f /etc/default/rcS ] && . /etc/default/rcS
+
+# Define LSB log_* functions.
+. /lib/lsb/init-functions
+
+do_start()
+{
+ # Return
+ # 0 if daemon has been started
+ # 1 if daemon was already running
+ # 2 if daemon could not be started
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON --test > /dev/null \
+ || return 1
+ start-stop-daemon --start --quiet --pidfile $PIDFILE --exec $DAEMON -- \
+ $DAEMON_ARGS \
+ || return 2
+}
+
+do_stop()
+{
+ # Return
+ # 0 if daemon has been stopped
+ # 1 if daemon was already stopped
+ # 2 if daemon could not be stopped
+ # other if a failure occurred
+ start-stop-daemon --stop --quiet --retry=TERM/30/KILL/5 --pidfile $PIDFILE --name $NAME
+ RETVAL="$?"
+ [ "$RETVAL" = 2 ] && return 2
+ start-stop-daemon --stop --quiet --oknodo --retry=0/30/KILL/5 --exec $DAEMON
+ [ "$?" = 2 ] && return 2
+ rm -f $PIDFILE
+ return "$RETVAL"
+}
+
+do_reload() {
+ start-stop-daemon --stop --signal HUP --quiet --pidfile $PIDFILE --name $NAME
+ return 0
+}
+
+case "$1" in
+ start)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Starting $DESC" "$NAME"
+ do_start
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ stop)
+ [ "$VERBOSE" != no ] && log_daemon_msg "Stopping $DESC" "$NAME"
+ do_stop
+ case "$?" in
+ 0|1) [ "$VERBOSE" != no ] && log_end_msg 0 ;;
+ 2) [ "$VERBOSE" != no ] && log_end_msg 1 ;;
+ esac
+ ;;
+ reload|force-reload)
+ log_daemon_msg "Reloading $DESC" "$NAME"
+ do_reload
+ log_end_msg $?
+ ;;
+ restart)
+ log_daemon_msg "Restarting $DESC" "$NAME"
+ do_stop
+ sleep 1
+ case "$?" in
+ 0|1)
+ do_start
+ case "$?" in
+ 0) log_end_msg 0 ;;
+ 1) log_end_msg 1 ;; # Old process is still running
+ *) log_end_msg 1 ;; # Failed to start
+ esac
+ ;;
+ *)
+ # Failed to stop
+ log_end_msg 1
+ ;;
+ esac
+ ;;
+ *)
+ echo "Usage: $SCRIPTNAME {start|stop|restart|reload|force-reload}" >&2
+ exit 3
+ ;;
+esac
+
+:
- change mode from '' to '0755'
- change owner from '' to 'root'
- change group from '' to 'root'
* service[pbs_sched] action enable
- enable service service[pbs_sched]
* service[pbs_sched] action start
- start service service[pbs_sched]
* template[/opt/cfncluster/scripts/publish_pending] action create
- create new file /opt/cfncluster/scripts/publish_pending
- update content in file /opt/cfncluster/scripts/publish_pending from none to 7b219d
--- /opt/cfncluster/scripts/publish_pending 2016-05-27 04:04:50.714866999 +0000
+++ /opt/cfncluster/scripts/.chef-publish_pending20160527-1708-jf7ptu 2016-05-27 04:04:50.714866999 +0000
@@ -1 +1,27 @@
+#!/bin/bash
+
+# Copyright 2013-2016 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+#
+# Licensed under the Amazon Software License (the "License"). You may not use this file except in compliance with the
+# License. A copy of the License is located at
+#
+# http://aws.amazon.com/asl/
+#
+# or in the "LICENSE.txt" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES
+# OR CONDITIONS OF ANY KIND, express or implied. See the License for the specific language governing permissions and
+# limitations under the License.
+
+PATH=/bin:/usr/bin:/usr/local/bin
+export PATH
+
+. /opt/cfncluster/cfnconfig
+
+. /etc/profile.d/torque.sh
+pending=$(qstat -i -t | tail -n+6 | awk '{total = total+ $7}END{print total}')
+
+if [ "${pending}x" == "x" ]; then
+pending=0
+fi
+
+aws --region ${cfn_region} cloudwatch put-metric-data --namespace cfncluster --metric-name pending --unit Count --value ${pending} --dimensions Stack=${stack_name}
- change mode from '' to '0744'
- change owner from '' to 'root'
- change group from '' to 'root'
* cron[publish_pending] action create
- add crontab entry for cron[publish_pending]
Recipe: openssh::default
* service[ssh] action restart
- restart service service[ssh]
Recipe: nfs::server
* service[nfs-kernel-server] action restart
- restart service service[nfs-kernel-server]
Running handlers:
Running handlers complete
Chef Client finished, 71/202 resources updated in 02 minutes 13 seconds
2016-05-27 04:04:52,361 [DEBUG] No services specified
2016-05-27 04:04:52,363 [INFO] Running config shellRunPostInstall
2016-05-27 04:04:52,363 [DEBUG] No packages specified
2016-05-27 04:04:52,363 [DEBUG] No groups specified
2016-05-27 04:04:52,363 [DEBUG] No users specified
2016-05-27 04:04:52,363 [DEBUG] No sources specified
2016-05-27 04:04:52,363 [DEBUG] No files specified
2016-05-27 04:04:52,363 [DEBUG] Running command runpostinstall
2016-05-27 04:04:52,364 [DEBUG] No test for command runpostinstall
2016-05-27 04:04:52,658 [ERROR] Command runpostinstall (/opt/cfncluster/scripts/fetch_and_run -postinstall) failed
2016-05-27 04:04:52,658 [DEBUG] Command runpostinstall output: Warning: unknown mime-type for "hi!" -- using "application/octet-stream"
Error: no such file "hi!"
cfncluster: fetch_and_run - Failed to run boot_as_master postinstall
2016-05-27 04:04:52,658 [ERROR] Error encountered during build of shellRunPostInstall: Command runpostinstall failed
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 517, in run_config
CloudFormationCarpenter(config, self._auth_config).build(worklog)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 248, in build
changes['commands'] = CommandTool().apply(self._config.commands)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/command_tool.py", line 113, in apply
raise ToolError(u"Command %s failed" % name)
ToolError: Command runpostinstall failed
2016-05-27 04:04:52,659 [ERROR] -----------------------BUILD FAILED!------------------------
2016-05-27 04:04:52,661 [ERROR] Unhandled exception during build: Command runpostinstall failed
Traceback (most recent call last):
File "/usr/local/bin/cfn-init", line 123, in <module>
worklog.build(detail.metadata, configSets)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 118, in build
Contractor(metadata).build(configSets, self)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 505, in build
self.run_config(config, worklog)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 517, in run_config
CloudFormationCarpenter(config, self._auth_config).build(worklog)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/construction.py", line 248, in build
changes['commands'] = CommandTool().apply(self._config.commands)
File "/usr/local/lib/python2.7/dist-packages/cfnbootstrap/command_tool.py", line 113, in apply
raise ToolError(u"Command %s failed" % name)
ToolError: Command runpostinstall failed
2016-05-27 04:04:53,035 [DEBUG] CloudFormation client initialized with endpoint https://cloudformation.us-west-2.amazonaws.com
2016-05-27 04:04:53,035 [DEBUG] Signaling resource MasterServer in stack cfncluster-pepper with unique ID i-cfef1014 and status FAILURE
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment