-
-
Save tomfaulhaber/11262826 to your computer and use it in GitHub Desktop.
bash_profile not built
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Local: master ~/src/clj/pallet/hadoop-crate/ | |
Remote: master @ origin (https://github.com/palletops/hadoop-crate.git) | |
Head: ea401ff Merge release/0.1.6 | |
Tag: 0.1.6 (2 ahead) | |
Unstaged changes: | |
Modified src/palletops/crate/hadoop/base.clj | |
diff --git a/src/palletops/crate/hadoop/base.clj b/src/palletops/crate/hadoop/base.clj | |
index f0fca38..0a20b35 100644 | |
--- a/src/palletops/crate/hadoop/base.clj | |
+++ b/src/palletops/crate/hadoop/base.clj | |
@@ -337,8 +337,9 @@ kernel.* Properties | |
(defplan hadoop-user | |
"Create the hadoop user, setting java home and path" | |
[{:keys [instance-id] :as options}] | |
- (let [{:keys [user group home]} | |
- (get-settings :hadoop {:instance-id instance-id})] | |
+ (let [{:keys [user group home] :as settings} | |
+ (get-settings :hadoop {:instance-id instance-id}) | |
+ hadoop-dir (default-home settings)] | |
(group-action group :system true) | |
(user-action user :group group :system true :create-home true :shell :bash) | |
(remote-file | |
@@ -348,7 +349,7 @@ kernel.* Properties | |
:literal true | |
:content (script | |
(set! JAVA_HOME (~java-home)) | |
- (set! PATH (str @PATH ":" ~home "/bin")))))) | |
+ (set! PATH (str @PATH ":" ~home "/bin:" ~hadoop-dir "/bin")))))) | |
;;; # Property files | |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
; CIDER 0.6.0alpha (package: 20140125.2304) (Clojure 1.4.0, nREPL 0.2.3) | |
user> | |
#<Namespace hadoop-int.hadoop> | |
hadoop-int.hadoop> | |
hadoop-int.hadoop> (def c (start-cluster)) | |
#'hadoop-int.hadoop/c | |
hadoop-int.hadoop> (pprint @c) | |
{:initial-plan-state {:pallet.action/action-options {}}, | |
:environment | |
{:compute | |
#<JcloudsService pallet.compute.jclouds.JcloudsService@20737c37>, | |
:user | |
{:username "tom", | |
:public-key-path "/Users/tom/.ssh/id_rsa.pub", | |
:private-key-path "/Users/tom/.ssh/id_rsa", | |
:public-key nil, | |
:private-key nil, | |
:passphrase nil, | |
:password nil, | |
:sudo-password nil, | |
:no-sudo nil, | |
:sudo-user nil, | |
:state-root nil, | |
:state-group nil}}, | |
:new-nodes | |
({:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]} | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}), | |
:old-nodes (), | |
:targets | |
({:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]} | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}), | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy :palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode ["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy :palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class "org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode ["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:service-state | |
({:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]} | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}), | |
:results | |
({:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-da664189" | |
{:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"{\n:os \"Linux\"\n:rev \"3.2.0-60-virtual\"\n:mach \"x86_64\"\n}\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho {\necho :os '\"'$(uname -s)'\"'\necho :rev '\"'$(uname -r)'\"'\necho :mach '\"'$(uname -m)'\"'\necho }\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "os: infer-os: "} | |
{:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :linux, | |
:os-version "3.2.0-60-virtual", | |
:action-symbol clj-action12976, | |
:context "os: infer-os: "} | |
{:exit 0, | |
:out "{\n:id \"Ubuntu\"\n:release \"12.04\"\n}\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif [ -e /etc/debconf_version ]; then\nID=$(cat /etc/redhat-release | egrep -o -e '^[A-Za-z ]+release' | sed -e 's/ release//')\nRELEASE=$(lsb_release -s -r)\nfi\nif [ -e /etc/lsb-release ]; then\nsource /etc/lsb-release\nID=${DISTRIB_ID}\nRELEASE=${DISTRIB_RELEASE}\nfi\nif [ -e /etc/redhat-release ]; then\nID=$(cat /etc/redhat-release | egrep -o -e '^[A-Za-z ]+release' | sed -e 's/ release//')\nRELEASE=$(cat /etc/redhat-release | sed -e 's/.*release//' | sed -e 's/[^0-9.]//g')\nfi\nif [ -e /etc/SUSE-release ]; then\nID=$(cat /etc/SUSE-release | tr ' ' | sed -e 's/VERSION.*//')\nRELEASE=$(cat /etc/SUSE-release | tr ' ' | sed -e 's/.*= //')\nfi\nif [ -e /etc/mandrake-release ]; then\nID=Mandrake\nRELEASE=$(cat /etc/mandrake-release | sed -e 's/.*release //' | sed -e 's/ .*//')\nfi\necho {\necho :id '\"'${ID:-unknown}'\"'\necho :release '\"'${RELEASE:-unknown}'\"'\necho }\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "os: infer-distro: "} | |
{:release "12.04", | |
:id "Ubuntu", | |
:os-family :ubuntu, | |
:os-version "12.04", | |
:action-symbol clj-action13001, | |
:context "os: infer-distro: "} | |
{:os-version "12.04", | |
:os-family :ubuntu, | |
:os "Linux", | |
:rev "3.2.0-60-virtual", | |
:mach "x86_64", | |
:release "12.04", | |
:id "Ubuntu", | |
:action-symbol clj-action13025, | |
:context "os: "} | |
{:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04", | |
:action-symbol clj-action13025, | |
:context "os: "}), | |
:phase :pallet/os} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"{\n:os \"Linux\"\n:rev \"3.2.0-60-virtual\"\n:mach \"x86_64\"\n}\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho {\necho :os '\"'$(uname -s)'\"'\necho :rev '\"'$(uname -r)'\"'\necho :mach '\"'$(uname -m)'\"'\necho }\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "os: infer-os: "} | |
{:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :linux, | |
:os-version "3.2.0-60-virtual", | |
:action-symbol clj-action12976, | |
:context "os: infer-os: "} | |
{:exit 0, | |
:out "{\n:id \"Ubuntu\"\n:release \"12.04\"\n}\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif [ -e /etc/debconf_version ]; then\nID=$(cat /etc/redhat-release | egrep -o -e '^[A-Za-z ]+release' | sed -e 's/ release//')\nRELEASE=$(lsb_release -s -r)\nfi\nif [ -e /etc/lsb-release ]; then\nsource /etc/lsb-release\nID=${DISTRIB_ID}\nRELEASE=${DISTRIB_RELEASE}\nfi\nif [ -e /etc/redhat-release ]; then\nID=$(cat /etc/redhat-release | egrep -o -e '^[A-Za-z ]+release' | sed -e 's/ release//')\nRELEASE=$(cat /etc/redhat-release | sed -e 's/.*release//' | sed -e 's/[^0-9.]//g')\nfi\nif [ -e /etc/SUSE-release ]; then\nID=$(cat /etc/SUSE-release | tr ' ' | sed -e 's/VERSION.*//')\nRELEASE=$(cat /etc/SUSE-release | tr ' ' | sed -e 's/.*= //')\nfi\nif [ -e /etc/mandrake-release ]; then\nID=Mandrake\nRELEASE=$(cat /etc/mandrake-release | sed -e 's/.*release //' | sed -e 's/ .*//')\nfi\necho {\necho :id '\"'${ID:-unknown}'\"'\necho :release '\"'${RELEASE:-unknown}'\"'\necho }\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "os: infer-distro: "} | |
{:release "12.04", | |
:id "Ubuntu", | |
:os-family :ubuntu, | |
:os-version "12.04", | |
:action-symbol clj-action13001, | |
:context "os: infer-distro: "} | |
{:os-version "12.04", | |
:os-family :ubuntu, | |
:os "Linux", | |
:rev "3.2.0-60-virtual", | |
:mach "x86_64", | |
:release "12.04", | |
:id "Ubuntu", | |
:action-symbol clj-action13025, | |
:context "os: "} | |
{:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04", | |
:action-symbol clj-action13025, | |
:context "os: "}), | |
:phase :pallet/os} | |
{:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result (), | |
:phase :settings} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result (), | |
:phase :settings} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"package-manager update ...\n#> package-manager update : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'package-manager update ...';\n{\napt-get -qq update\n } || { echo '#> package-manager update : FAIL'; exit 1;} >&2 \necho '#> package-manager update : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package-manager, | |
:context nil} | |
{:exit 0, | |
:out | |
"[automated-admin-user: install]: Packages...\nReading package lists...\nBuilding dependency tree...\nReading state information...\nsudo is already the newest version.\n0 upgraded, 0 newly installed, 0 to remove and 8 not upgraded.\naccountsservice\t\t\t\t\tinstall\nacpid\t\t\t\t\t\tinstall\nadduser\t\t\t\t\t\tinstall\napparmor\t\t\t\t\tinstall\napport\t\t\t\t\t\tinstall\napport-symptoms\t\t\t\t\tinstall\napt\t\t\t\t\t\tinstall\napt-transport-https\t\t\t\tinstall\napt-utils\t\t\t\t\tinstall\napt-xapian-index\t\t\t\tinstall\naptitude\t\t\t\t\tinstall\nat\t\t\t\t\t\tinstall\nbase-files\t\t\t\t\tinstall\nbase-passwd\t\t\t\t\tinstall\nbash\t\t\t\t\t\tinstall\nbash-completion\t\t\t\t\tinstall\nbc\t\t\t\t\t\tinstall\nbind9-host\t\t\t\t\tinstall\nbsdmainutils\t\t\t\t\tinstall\nbsdutils\t\t\t\t\tinstall\nbusybox-initramfs\t\t\t\tinstall\nbusybox-static\t\t\t\t\tinstall\nbyobu\t\t\t\t\t\tinstall\nbzip2\t\t\t\t\t\tinstall\nca-certificates\t\t\t\t\tinstall\ncloud-init\t\t\t\t\tinstall\ncloud-initramfs-growroot\t\t\tinstall\ncloud-initramfs-rescuevol\t\t\tinstall\ncloud-utils\t\t\t\t\tinstall\ncommand-not-found\t\t\t\tinstall\ncommand-not-found-data\t\t\t\tinstall\nconsole-setup\t\t\t\t\tinstall\ncoreutils\t\t\t\t\tinstall\ncpio\t\t\t\t\t\tinstall\ncrda\t\t\t\t\t\tinstall\ncron\t\t\t\t\t\tinstall\ncurl\t\t\t\t\t\tinstall\ndash\t\t\t\t\t\tinstall\ndbus\t\t\t\t\t\tinstall\ndebconf\t\t\t\t\t\tinstall\ndebconf-i18n\t\t\t\t\tinstall\ndebianutils\t\t\t\t\tinstall\ndiffutils\t\t\t\t\tinstall\ndmidecode\t\t\t\t\tinstall\ndmsetup\t\t\t\t\t\tinstall\ndnsutils\t\t\t\t\tinstall\ndosfstools\t\t\t\t\tinstall\ndpkg\t\t\t\t\t\tinstall\ne2fslibs\t\t\t\t\tinstall\ne2fsprogs\t\t\t\t\tinstall\ned\t\t\t\t\t\tinstall\neject\t\t\t\t\t\tinstall\neuca2ools\t\t\t\t\tinstall\nfile\t\t\t\t\t\tinstall\nfindutils\t\t\t\t\tinstall\nfonts-ubuntu-font-family-console\t\tinstall\nfriendly-recovery\t\t\t\tinstall\nftp\t\t\t\t\t\tinstall\nfuse\t\t\t\t\t\tinstall\ngcc-4.6-base\t\t\t\t\tinstall\ngeoip-database\t\t\t\t\tinstall\ngettext-base\t\t\t\t\tinstall\ngir1.2-glib-2.0\t\t\t\t\tinstall\ngnupg\t\t\t\t\t\tinstall\ngpgv\t\t\t\t\t\tinstall\ngrep\t\t\t\t\t\tinstall\ngroff-base\t\t\t\t\tinstall\ngrub-common\t\t\t\t\tinstall\ngrub-gfxpayload-lists\t\t\t\tinstall\ngrub-legacy-ec2\t\t\t\t\tinstall\ngrub-pc\t\t\t\t\t\tinstall\ngrub-pc-bin\t\t\t\t\tinstall\ngrub2-common\t\t\t\t\tinstall\ngzip\t\t\t\t\t\tinstall\nhdparm\t\t\t\t\t\tinstall\nhostname\t\t\t\t\tinstall\nifupdown\t\t\t\t\tinstall\ninfo\t\t\t\t\t\tinstall\ninitramfs-tools\t\t\t\t\tinstall\ninitramfs-tools-bin\t\t\t\tinstall\ninitscripts\t\t\t\t\tinstall\ninsserv\t\t\t\t\t\tinstall\ninstall-info\t\t\t\t\tinstall\niproute\t\t\t\t\t\tinstall\niptables\t\t\t\t\tinstall\niputils-ping\t\t\t\t\tinstall\niputils-tracepath\t\t\t\tinstall\nirqbalance\t\t\t\t\tinstall\nisc-dhcp-client\t\t\t\t\tinstall\nisc-dhcp-common\t\t\t\t\tinstall\niso-codes\t\t\t\t\tinstall\niw\t\t\t\t\t\tinstall\nkbd\t\t\t\t\t\tinstall\nkeyboard-configuration\t\t\t\tinstall\nklibc-utils\t\t\t\t\tinstall\nkrb5-locales\t\t\t\t\tinstall\nlandscape-client\t\t\t\tinstall\nlandscape-common\t\t\t\tinstall\nlanguage-selector-common\t\t\tinstall\nlaptop-detect\t\t\t\t\tinstall\nless\t\t\t\t\t\tinstall\nlibaccountsservice0\t\t\t\tinstall\nlibacl1\t\t\t\t\t\tinstall\nlibapt-inst1.4\t\t\t\t\tinstall\nlibapt-pkg4.12\t\t\t\t\tinstall\nlibasn1-8-heimdal\t\t\t\tinstall\nlibattr1\t\t\t\t\tinstall\nlibbind9-80\t\t\t\t\tinstall\nlibblkid1\t\t\t\t\tinstall\nlibboost-iostreams1.46.1\t\t\tinstall\nlibbsd0\t\t\t\t\t\tinstall\nlibbz2-1.0\t\t\t\t\tinstall\nlibc-bin\t\t\t\t\tinstall\nlibc6\t\t\t\t\t\tinstall\nlibcap-ng0\t\t\t\t\tinstall\nlibclass-accessor-perl\t\t\t\tinstall\nlibclass-isa-perl\t\t\t\tinstall\nlibcomerr2\t\t\t\t\tinstall\nlibcurl3\t\t\t\t\tinstall\nlibcurl3-gnutls\t\t\t\t\tinstall\nlibcwidget3\t\t\t\t\tinstall\nlibdb5.1\t\t\t\t\tinstall\nlibdbus-1-3\t\t\t\t\tinstall\nlibdbus-glib-1-2\t\t\t\tinstall\nlibdevmapper1.02.1\t\t\t\tinstall\nlibdns81\t\t\t\t\tinstall\nlibdrm-intel1\t\t\t\t\tinstall\nlibdrm-nouveau1a\t\t\t\tinstall\nlibdrm-radeon1\t\t\t\t\tinstall\nlibdrm2\t\t\t\t\t\tinstall\nlibedit2\t\t\t\t\tinstall\nlibelf1\t\t\t\t\t\tinstall\nlibept1.4.12\t\t\t\t\tinstall\nlibevent-2.0-5\t\t\t\t\tinstall\nlibexpat1\t\t\t\t\tinstall\nlibffi6\t\t\t\t\t\tinstall\nlibfreetype6\t\t\t\t\tinstall\nlibfribidi0\t\t\t\t\tinstall\nlibfuse2\t\t\t\t\tinstall\nlibgc1c2\t\t\t\t\tinstall\nlibgcc1\t\t\t\t\t\tinstall\nlibgcrypt11\t\t\t\t\tinstall\nlibgdbm3\t\t\t\t\tinstall\nlibgeoip1\t\t\t\t\tinstall\nlibgirepository-1.0-1\t\t\t\tinstall\nlibglib2.0-0\t\t\t\t\tinstall\nlibgmp10\t\t\t\t\tinstall\nlibgnutls26\t\t\t\t\tinstall\nlibgpg-error0\t\t\t\t\tinstall\nlibgpm2\t\t\t\t\t\tinstall\nlibgssapi-krb5-2\t\t\t\tinstall\nlibgssapi3-heimdal\t\t\t\tinstall\nlibhcrypto4-heimdal\t\t\t\tinstall\nlibheimbase1-heimdal\t\t\t\tinstall\nlibheimntlm0-heimdal\t\t\t\tinstall\nlibhx509-5-heimdal\t\t\t\tinstall\nlibidn11\t\t\t\t\tinstall\nlibio-string-perl\t\t\t\tinstall\nlibisc83\t\t\t\t\tinstall\nlibisccc80\t\t\t\t\tinstall\nlibisccfg82\t\t\t\t\tinstall\nlibiw30\t\t\t\t\t\tinstall\nlibjs-jquery\t\t\t\t\tinstall\nlibk5crypto3\t\t\t\t\tinstall\nlibkeyutils1\t\t\t\t\tinstall\nlibklibc\t\t\t\t\tinstall\nlibkrb5-26-heimdal\t\t\t\tinstall\nlibkrb5-3\t\t\t\t\tinstall\nlibkrb5support0\t\t\t\t\tinstall\nlibldap-2.4-2\t\t\t\t\tinstall\nliblocale-gettext-perl\t\t\t\tinstall\nliblockfile-bin\t\t\t\t\tinstall\nliblockfile1\t\t\t\t\tinstall\nliblwres80\t\t\t\t\tinstall\nliblzma5\t\t\t\t\tinstall\nlibmagic1\t\t\t\t\tinstall\nlibmount1\t\t\t\t\tinstall\nlibncurses5\t\t\t\t\tinstall\nlibncursesw5\t\t\t\t\tinstall\nlibnewt0.52\t\t\t\t\tinstall\nlibnfnetlink0\t\t\t\t\tinstall\nlibnih-dbus1\t\t\t\t\tinstall\nlibnih1\t\t\t\t\t\tinstall\nlibnl-3-200\t\t\t\t\tinstall\nlibnl-genl-3-200\t\t\t\tinstall\nlibp11-kit0\t\t\t\t\tinstall\nlibpam-modules\t\t\t\t\tinstall\nlibpam-modules-bin\t\t\t\tinstall\nlibpam-runtime\t\t\t\t\tinstall\nlibpam0g\t\t\t\t\tinstall\nlibparse-debianchangelog-perl\t\t\tinstall\nlibparted0debian1\t\t\t\tinstall\nlibpcap0.8\t\t\t\t\tinstall\nlibpci3\t\t\t\t\t\tinstall\nlibpciaccess0\t\t\t\t\tinstall\nlibpcre3\t\t\t\t\tinstall\nlibpcsclite1\t\t\t\t\tinstall\nlibpipeline1\t\t\t\t\tinstall\nlibplymouth2\t\t\t\t\tinstall\nlibpng12-0\t\t\t\t\tinstall\nlibpolkit-gobject-1-0\t\t\t\tinstall\nlibpopt0\t\t\t\t\tinstall\nlibpython2.7\t\t\t\t\tinstall\nlibreadline6\t\t\t\t\tinstall\nlibroken18-heimdal\t\t\t\tinstall\nlibrtmp0\t\t\t\t\tinstall\nlibsasl2-2\t\t\t\t\tinstall\nlibsasl2-modules\t\t\t\tinstall\nlibselinux1\t\t\t\t\tinstall\nlibsigc++-2.0-0c2a\t\t\t\tinstall\nlibslang2\t\t\t\t\tinstall\nlibsqlite3-0\t\t\t\t\tinstall\nlibss2\t\t\t\t\t\tinstall\nlibssl1.0.0\t\t\t\t\tinstall\nlibstdc++6\t\t\t\t\tinstall\nlibsub-name-perl\t\t\t\tinstall\nlibswitch-perl\t\t\t\t\tinstall\nlibtasn1-3\t\t\t\t\tinstall\nlibtext-charwidth-perl\t\t\t\tinstall\nlibtext-iconv-perl\t\t\t\tinstall\nlibtext-wrapi18n-perl\t\t\t\tinstall\nlibtimedate-perl\t\t\t\tinstall\nlibtinfo5\t\t\t\t\tinstall\nlibudev0\t\t\t\t\tinstall\nlibusb-0.1-4\t\t\t\t\tinstall\nlibusb-1.0-0\t\t\t\t\tinstall\nlibuuid1\t\t\t\t\tinstall\nlibwind0-heimdal\t\t\t\tinstall\nlibwrap0\t\t\t\t\tinstall\nlibx11-6\t\t\t\t\tinstall\nlibx11-data\t\t\t\t\tinstall\nlibxapian22\t\t\t\t\tinstall\nlibxau6\t\t\t\t\t\tinstall\nlibxcb1\t\t\t\t\t\tinstall\nlibxdmcp6\t\t\t\t\tinstall\nlibxext6\t\t\t\t\tinstall\nlibxml2\t\t\t\t\t\tinstall\nlibxmuu1\t\t\t\t\tinstall\nlibyaml-0-2\t\t\t\t\tinstall\nlinux-firmware\t\t\t\t\tinstall\nlinux-headers-3.2.0-60\t\t\t\tinstall\nlinux-headers-3.2.0-60-virtual\t\t\tinstall\nlinux-headers-virtual\t\t\t\tinstall\nlinux-image-3.2.0-60-virtual\t\t\tinstall\nlinux-image-virtual\t\t\t\tinstall\nlinux-virtual\t\t\t\t\tinstall\nlocales\t\t\t\t\t\tinstall\nlockfile-progs\t\t\t\t\tinstall\nlogin\t\t\t\t\t\tinstall\nlogrotate\t\t\t\t\tinstall\nlsb-base\t\t\t\t\tinstall\nlsb-release\t\t\t\t\tinstall\nlshw\t\t\t\t\t\tinstall\nlsof\t\t\t\t\t\tinstall\nltrace\t\t\t\t\t\tinstall\nmakedev\t\t\t\t\t\tinstall\nman-db\t\t\t\t\t\tinstall\nmanpages\t\t\t\t\tinstall\nmawk\t\t\t\t\t\tinstall\nmemtest86+\t\t\t\t\tinstall\nmime-support\t\t\t\t\tinstall\nmlocate\t\t\t\t\t\tinstall\nmodule-init-tools\t\t\t\tinstall\nmount\t\t\t\t\t\tinstall\nmountall\t\t\t\t\tinstall\nmtr-tiny\t\t\t\t\tinstall\nmultiarch-support\t\t\t\tinstall\nnano\t\t\t\t\t\tinstall\nncurses-base\t\t\t\t\tinstall\nncurses-bin\t\t\t\t\tinstall\nnet-tools\t\t\t\t\tinstall\nnetbase\t\t\t\t\t\tinstall\nnetcat-openbsd\t\t\t\t\tinstall\nntfs-3g\t\t\t\t\t\tinstall\nntpdate\t\t\t\t\t\tinstall\nopenssh-client\t\t\t\t\tinstall\nopenssh-server\t\t\t\t\tinstall\nopenssl\t\t\t\t\t\tinstall\nos-prober\t\t\t\t\tinstall\nparted\t\t\t\t\t\tinstall\npasswd\t\t\t\t\t\tinstall\npatch\t\t\t\t\t\tinstall\npciutils\t\t\t\t\tinstall\nperl\t\t\t\t\t\tinstall\nperl-base\t\t\t\t\tinstall\nperl-modules\t\t\t\t\tinstall\nplymouth\t\t\t\t\tinstall\nplymouth-theme-ubuntu-text\t\t\tinstall\npopularity-contest\t\t\t\tinstall\npowermgmt-base\t\t\t\t\tinstall\nppp\t\t\t\t\t\tinstall\npppconfig\t\t\t\t\tinstall\npppoeconf\t\t\t\t\tinstall\nprocps\t\t\t\t\t\tinstall\npsmisc\t\t\t\t\t\tinstall\npython\t\t\t\t\t\tinstall\npython-apport\t\t\t\t\tinstall\npython-apt\t\t\t\t\tinstall\npython-apt-common\t\t\t\tinstall\npython-boto\t\t\t\t\tinstall\npython-chardet\t\t\t\t\tinstall\npython-cheetah\t\t\t\t\tinstall\npython-configobj\t\t\t\tinstall\npython-crypto\t\t\t\t\tinstall\npython-dbus\t\t\t\t\tinstall\npython-dbus-dev\t\t\t\t\tinstall\npython-debian\t\t\t\t\tinstall\npython-gdbm\t\t\t\t\tinstall\npython-gi\t\t\t\t\tinstall\npython-gnupginterface\t\t\t\tinstall\npython-httplib2\t\t\t\t\tinstall\npython-keyring\t\t\t\t\tinstall\npython-launchpadlib\t\t\t\tinstall\npython-lazr.restfulclient\t\t\tinstall\npython-lazr.uri\t\t\t\t\tinstall\npython-m2crypto\t\t\t\t\tinstall\npython-minimal\t\t\t\t\tinstall\npython-newt\t\t\t\t\tinstall\npython-oauth\t\t\t\t\tinstall\npython-openssl\t\t\t\t\tinstall\npython-pam\t\t\t\t\tinstall\npython-paramiko\t\t\t\t\tinstall\npython-pkg-resources\t\t\t\tinstall\npython-problem-report\t\t\t\tinstall\npython-pycurl\t\t\t\t\tinstall\npython-serial\t\t\t\t\tinstall\npython-simplejson\t\t\t\tinstall\npython-software-properties\t\t\tinstall\npython-twisted-bin\t\t\t\tinstall\npython-twisted-core\t\t\t\tinstall\npython-twisted-names\t\t\t\tinstall\npython-twisted-web\t\t\t\tinstall\npython-wadllib\t\t\t\t\tinstall\npython-xapian\t\t\t\t\tinstall\npython-yaml\t\t\t\t\tinstall\npython-zope.interface\t\t\t\tinstall\npython2.7\t\t\t\t\tinstall\npython2.7-minimal\t\t\t\tinstall\nreadline-common\t\t\t\t\tinstall\nresolvconf\t\t\t\t\tinstall\nrsync\t\t\t\t\t\tinstall\nrsyslog\t\t\t\t\t\tinstall\nscreen\t\t\t\t\t\tinstall\nsed\t\t\t\t\t\tinstall\nsensible-utils\t\t\t\t\tinstall\nsgml-base\t\t\t\t\tinstall\nssh-import-id\t\t\t\t\tinstall\nstrace\t\t\t\t\t\tinstall\nsudo\t\t\t\t\t\tinstall\nsysv-rc\t\t\t\t\t\tinstall\nsysvinit-utils\t\t\t\t\tinstall\ntar\t\t\t\t\t\tinstall\ntasksel\t\t\t\t\t\tinstall\ntasksel-data\t\t\t\t\tinstall\ntcpd\t\t\t\t\t\tinstall\ntcpdump\t\t\t\t\t\tinstall\ntelnet\t\t\t\t\t\tinstall\ntime\t\t\t\t\t\tinstall\ntmux\t\t\t\t\t\tinstall\ntzdata\t\t\t\t\t\tinstall\nubuntu-keyring\t\t\t\t\tinstall\nubuntu-minimal\t\t\t\t\tinstall\nubuntu-standard\t\t\t\t\tinstall\nucf\t\t\t\t\t\tinstall\nudev\t\t\t\t\t\tinstall\nufw\t\t\t\t\t\tinstall\nunattended-upgrades\t\t\t\tinstall\nupdate-manager-core\t\t\t\tinstall\nupdate-notifier-common\t\t\t\tinstall\nupstart\t\t\t\t\t\tinstall\nureadahead\t\t\t\t\tinstall\nusbutils\t\t\t\t\tinstall\nutil-linux\t\t\t\t\tinstall\nuuid-runtime\t\t\t\t\tinstall\nvim\t\t\t\t\t\tinstall\nvim-common\t\t\t\t\tinstall\nvim-runtime\t\t\t\t\tinstall\nvim-tiny\t\t\t\t\tinstall\nw3m\t\t\t\t\t\tinstall\nwget\t\t\t\t\t\tinstall\nwhiptail\t\t\t\t\tinstall\nwhoopsie\t\t\t\t\tinstall\nwireless-regdb\t\t\t\t\tinstall\nwireless-tools\t\t\t\t\tinstall\nwpasupplicant\t\t\t\t\tinstall\nxauth\t\t\t\t\t\tinstall\nxkb-data\t\t\t\t\tinstall\nxml-core\t\t\t\t\tinstall\nxz-lzma\t\t\t\t\t\tinstall\nxz-utils\t\t\t\t\tinstall\nzlib1g\t\t\t\t\t\tinstall\n#> [automated-admin-user: install]: Packages : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho '[automated-admin-user: install]: Packages...';\n{\n{ debconf-set-selections <<EOF\ndebconf debconf/frontend select noninteractive\ndebconf debconf/frontend seen false\nEOF\n} && enableStart() {\nrm /usr/sbin/policy-rc.d\n} && apt-get -q -y install sudo+ && dpkg --get-selections\n } || { echo '#> [automated-admin-user: install]: Packages : FAIL'; exit 1;} >&2 \necho '#> [automated-admin-user: install]: Packages : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package, | |
:context "[automated-admin-user: install]: "} | |
{:exit 0, | |
:out "", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent passwd tom; then /usr/sbin/usermod --shell \"/bin/bash\" tom;else /usr/sbin/useradd --shell \"/bin/bash\" --create-home tom;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/user, | |
:context "[automated-admin-user]: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/...\n#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/...';\n{\nmkdir -m \"755\" -p $(getent passwd tom | cut -d: -f6)/.ssh/ && chown --recursive tom $(getent passwd tom | cut -d: -f6)/.ssh/ && chmod 755 $(getent passwd tom | cut -d: -f6)/.ssh/\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys...\n#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys...';\n{\ntouch $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && chown tom $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && chmod 644 $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/file, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32)...\n#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32)...';\n{\nauth_file=$(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && if ! ( fgrep \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDnV92GSukTdPwSitdUzox9hMlwYz3oApmJ9FY+7GQ/QNMfLD7jFDmn9wLqJU6WpWtXd5S+RfILZzS0LvPgFgWrY8/ZxcwwVlnp0adGdxb3s3iwsnSLFwF2KhyI6rpbDuFZaO+0ExbNUWC8YBXcp/9wQlsZ2OUJ/jQkpqwZ+L0sLsGS2cSH1lGwpYUr/oZ15gbSZ6mvqzBbZnZXI3IABd23b7xdaAldV3PYHCdJ6zSUNa675YsTj+SdEBPQpx0IakcH+XAp62ePchxMocEKeeTfW/aYDscxBw7M9aGZTCsNl4LKMzeF7FYlPAF1VYtXg31YxACKEpRHW+IVY9oSDenV tom@infolace.com\" ${auth_file} ); then\necho \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDnV92GSukTdPwSitdUzox9hMlwYz3oApmJ9FY+7GQ/QNMfLD7jFDmn9wLqJU6WpWtXd5S+RfILZzS0LvPgFgWrY8/ZxcwwVlnp0adGdxb3s3iwsnSLFwF2KhyI6rpbDuFZaO+0ExbNUWC8YBXcp/9wQlsZ2OUJ/jQkpqwZ+L0sLsGS2cSH1lGwpYUr/oZ15gbSZ6mvqzBbZnZXI3IABd23b7xdaAldV3PYHCdJ6zSUNa675YsTj+SdEBPQpx0IakcH+XAp62ePchxMocEKeeTfW/aYDscxBw7M9aGZTCsNl4LKMzeF7FYlPAF1VYtXg31YxACKEpRHW+IVY9oSDenV tom@infolace.com\n\" >> ${auth_file}\nfi\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37)...\n#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37)...';\n{\nif hash chcon 2>&- && [ -d /etc/selinux ] && [ -e /selinux/enforce ] && stat --format %C $(getent passwd tom | cut -d: -f6)/.ssh/ 2>&-; then chcon -Rv --type=user_home_t $(getent passwd tom | cut -d: -f6)/.ssh/;fi\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: sudoers: remote-file /etc/sudoers...\n--- /etc/sudoers\t2012-01-31 15:56:42.000000000 +0000\n+++ /tmp/root/bBCdbpTJf7UZAHwfBogFgg/sudoers\t2014-04-24 16:56:04.103442003 +0000\n@@ -1,29 +1,4 @@\n-#\n-# This file MUST be edited with the 'visudo' command as root.\n-#\n-# Please consider adding local content in /etc/sudoers.d/ instead of\n-# directly modifying this file.\n-#\n-# See the man page for details on how to write a sudoers file.\n-#\n-Defaults\tenv_reset\n-Defaults\tsecure_path=\"/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"\n-\n-# Host alias specification\n-\n-# User alias specification\n-\n-# Cmnd alias specification\n-\n-# User privilege specification\n-root\tALL=(ALL:ALL) ALL\n-\n-# Members of the admin group may gain root privileges\n-%admin ALL=(ALL) ALL\n-\n-# Allow members of group sudo to execute any command\n-%sudo\tALL=(ALL:ALL) ALL\n-\n-# See sudoers(5) for more information on \"#include\" directives:\n-\n-#includedir /etc/sudoers.d\n+Defaults env_keep=SSH_AUTH_SOCK\n+root ALL = (ALL) ALL\n+%adm ALL = (ALL) ALL\n+tom ALL = (ALL) NOPASSWD: ALL\nMD5 sum is 46780992b159bec006d30e7c37b65c1c sudoers\n#> automated-admin-user: sudoers: remote-file /etc/sudoers : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: sudoers: remote-file /etc/sudoers...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /etc/sudoers /var/lib/pallet/etc/sudoers /var/lib/pallet/etc/sudoers.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) <<EOFpallet\nDefaults env_keep=SSH_AUTH_SOCK\nroot ALL = (ALL) ALL\n%adm ALL = (ALL) ALL\ntom ALL = (ALL) NOPASSWD: ALL\nEOFpallet\n } && contentdiff= && if [ -e /etc/sudoers ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) ]; then\ndiff -u /etc/sudoers ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && chgrp $(id -ng root) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && chmod 0440 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) /etc/sudoers && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /etc/sudoers /var/lib/pallet/etc/sudoers\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /etc/sudoers /var/lib/pallet/etc/sudoers\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /etc/sudoers /var/lib/pallet/etc/sudoers.md5\n}\nfi\n } || { echo '#> automated-admin-user: sudoers: remote-file /etc/sudoers : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: sudoers: remote-file /etc/sudoers : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "automated-admin-user: sudoers: "}), | |
:phase :bootstrap} | |
{:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"package-manager update ...\n#> package-manager update : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'package-manager update ...';\n{\napt-get -qq update\n } || { echo '#> package-manager update : FAIL'; exit 1;} >&2 \necho '#> package-manager update : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package-manager, | |
:context nil} | |
{:exit 0, | |
:out | |
"[automated-admin-user: install]: Packages...\nReading package lists...\nBuilding dependency tree...\nReading state information...\nsudo is already the newest version.\n0 upgraded, 0 newly installed, 0 to remove and 8 not upgraded.\naccountsservice\t\t\t\t\tinstall\nacpid\t\t\t\t\t\tinstall\nadduser\t\t\t\t\t\tinstall\napparmor\t\t\t\t\tinstall\napport\t\t\t\t\t\tinstall\napport-symptoms\t\t\t\t\tinstall\napt\t\t\t\t\t\tinstall\napt-transport-https\t\t\t\tinstall\napt-utils\t\t\t\t\tinstall\napt-xapian-index\t\t\t\tinstall\naptitude\t\t\t\t\tinstall\nat\t\t\t\t\t\tinstall\nbase-files\t\t\t\t\tinstall\nbase-passwd\t\t\t\t\tinstall\nbash\t\t\t\t\t\tinstall\nbash-completion\t\t\t\t\tinstall\nbc\t\t\t\t\t\tinstall\nbind9-host\t\t\t\t\tinstall\nbsdmainutils\t\t\t\t\tinstall\nbsdutils\t\t\t\t\tinstall\nbusybox-initramfs\t\t\t\tinstall\nbusybox-static\t\t\t\t\tinstall\nbyobu\t\t\t\t\t\tinstall\nbzip2\t\t\t\t\t\tinstall\nca-certificates\t\t\t\t\tinstall\ncloud-init\t\t\t\t\tinstall\ncloud-initramfs-growroot\t\t\tinstall\ncloud-initramfs-rescuevol\t\t\tinstall\ncloud-utils\t\t\t\t\tinstall\ncommand-not-found\t\t\t\tinstall\ncommand-not-found-data\t\t\t\tinstall\nconsole-setup\t\t\t\t\tinstall\ncoreutils\t\t\t\t\tinstall\ncpio\t\t\t\t\t\tinstall\ncrda\t\t\t\t\t\tinstall\ncron\t\t\t\t\t\tinstall\ncurl\t\t\t\t\t\tinstall\ndash\t\t\t\t\t\tinstall\ndbus\t\t\t\t\t\tinstall\ndebconf\t\t\t\t\t\tinstall\ndebconf-i18n\t\t\t\t\tinstall\ndebianutils\t\t\t\t\tinstall\ndiffutils\t\t\t\t\tinstall\ndmidecode\t\t\t\t\tinstall\ndmsetup\t\t\t\t\t\tinstall\ndnsutils\t\t\t\t\tinstall\ndosfstools\t\t\t\t\tinstall\ndpkg\t\t\t\t\t\tinstall\ne2fslibs\t\t\t\t\tinstall\ne2fsprogs\t\t\t\t\tinstall\ned\t\t\t\t\t\tinstall\neject\t\t\t\t\t\tinstall\neuca2ools\t\t\t\t\tinstall\nfile\t\t\t\t\t\tinstall\nfindutils\t\t\t\t\tinstall\nfonts-ubuntu-font-family-console\t\tinstall\nfriendly-recovery\t\t\t\tinstall\nftp\t\t\t\t\t\tinstall\nfuse\t\t\t\t\t\tinstall\ngcc-4.6-base\t\t\t\t\tinstall\ngeoip-database\t\t\t\t\tinstall\ngettext-base\t\t\t\t\tinstall\ngir1.2-glib-2.0\t\t\t\t\tinstall\ngnupg\t\t\t\t\t\tinstall\ngpgv\t\t\t\t\t\tinstall\ngrep\t\t\t\t\t\tinstall\ngroff-base\t\t\t\t\tinstall\ngrub-common\t\t\t\t\tinstall\ngrub-gfxpayload-lists\t\t\t\tinstall\ngrub-legacy-ec2\t\t\t\t\tinstall\ngrub-pc\t\t\t\t\t\tinstall\ngrub-pc-bin\t\t\t\t\tinstall\ngrub2-common\t\t\t\t\tinstall\ngzip\t\t\t\t\t\tinstall\nhdparm\t\t\t\t\t\tinstall\nhostname\t\t\t\t\tinstall\nifupdown\t\t\t\t\tinstall\ninfo\t\t\t\t\t\tinstall\ninitramfs-tools\t\t\t\t\tinstall\ninitramfs-tools-bin\t\t\t\tinstall\ninitscripts\t\t\t\t\tinstall\ninsserv\t\t\t\t\t\tinstall\ninstall-info\t\t\t\t\tinstall\niproute\t\t\t\t\t\tinstall\niptables\t\t\t\t\tinstall\niputils-ping\t\t\t\t\tinstall\niputils-tracepath\t\t\t\tinstall\nirqbalance\t\t\t\t\tinstall\nisc-dhcp-client\t\t\t\t\tinstall\nisc-dhcp-common\t\t\t\t\tinstall\niso-codes\t\t\t\t\tinstall\niw\t\t\t\t\t\tinstall\nkbd\t\t\t\t\t\tinstall\nkeyboard-configuration\t\t\t\tinstall\nklibc-utils\t\t\t\t\tinstall\nkrb5-locales\t\t\t\t\tinstall\nlandscape-client\t\t\t\tinstall\nlandscape-common\t\t\t\tinstall\nlanguage-selector-common\t\t\tinstall\nlaptop-detect\t\t\t\t\tinstall\nless\t\t\t\t\t\tinstall\nlibaccountsservice0\t\t\t\tinstall\nlibacl1\t\t\t\t\t\tinstall\nlibapt-inst1.4\t\t\t\t\tinstall\nlibapt-pkg4.12\t\t\t\t\tinstall\nlibasn1-8-heimdal\t\t\t\tinstall\nlibattr1\t\t\t\t\tinstall\nlibbind9-80\t\t\t\t\tinstall\nlibblkid1\t\t\t\t\tinstall\nlibboost-iostreams1.46.1\t\t\tinstall\nlibbsd0\t\t\t\t\t\tinstall\nlibbz2-1.0\t\t\t\t\tinstall\nlibc-bin\t\t\t\t\tinstall\nlibc6\t\t\t\t\t\tinstall\nlibcap-ng0\t\t\t\t\tinstall\nlibclass-accessor-perl\t\t\t\tinstall\nlibclass-isa-perl\t\t\t\tinstall\nlibcomerr2\t\t\t\t\tinstall\nlibcurl3\t\t\t\t\tinstall\nlibcurl3-gnutls\t\t\t\t\tinstall\nlibcwidget3\t\t\t\t\tinstall\nlibdb5.1\t\t\t\t\tinstall\nlibdbus-1-3\t\t\t\t\tinstall\nlibdbus-glib-1-2\t\t\t\tinstall\nlibdevmapper1.02.1\t\t\t\tinstall\nlibdns81\t\t\t\t\tinstall\nlibdrm-intel1\t\t\t\t\tinstall\nlibdrm-nouveau1a\t\t\t\tinstall\nlibdrm-radeon1\t\t\t\t\tinstall\nlibdrm2\t\t\t\t\t\tinstall\nlibedit2\t\t\t\t\tinstall\nlibelf1\t\t\t\t\t\tinstall\nlibept1.4.12\t\t\t\t\tinstall\nlibevent-2.0-5\t\t\t\t\tinstall\nlibexpat1\t\t\t\t\tinstall\nlibffi6\t\t\t\t\t\tinstall\nlibfreetype6\t\t\t\t\tinstall\nlibfribidi0\t\t\t\t\tinstall\nlibfuse2\t\t\t\t\tinstall\nlibgc1c2\t\t\t\t\tinstall\nlibgcc1\t\t\t\t\t\tinstall\nlibgcrypt11\t\t\t\t\tinstall\nlibgdbm3\t\t\t\t\tinstall\nlibgeoip1\t\t\t\t\tinstall\nlibgirepository-1.0-1\t\t\t\tinstall\nlibglib2.0-0\t\t\t\t\tinstall\nlibgmp10\t\t\t\t\tinstall\nlibgnutls26\t\t\t\t\tinstall\nlibgpg-error0\t\t\t\t\tinstall\nlibgpm2\t\t\t\t\t\tinstall\nlibgssapi-krb5-2\t\t\t\tinstall\nlibgssapi3-heimdal\t\t\t\tinstall\nlibhcrypto4-heimdal\t\t\t\tinstall\nlibheimbase1-heimdal\t\t\t\tinstall\nlibheimntlm0-heimdal\t\t\t\tinstall\nlibhx509-5-heimdal\t\t\t\tinstall\nlibidn11\t\t\t\t\tinstall\nlibio-string-perl\t\t\t\tinstall\nlibisc83\t\t\t\t\tinstall\nlibisccc80\t\t\t\t\tinstall\nlibisccfg82\t\t\t\t\tinstall\nlibiw30\t\t\t\t\t\tinstall\nlibjs-jquery\t\t\t\t\tinstall\nlibk5crypto3\t\t\t\t\tinstall\nlibkeyutils1\t\t\t\t\tinstall\nlibklibc\t\t\t\t\tinstall\nlibkrb5-26-heimdal\t\t\t\tinstall\nlibkrb5-3\t\t\t\t\tinstall\nlibkrb5support0\t\t\t\t\tinstall\nlibldap-2.4-2\t\t\t\t\tinstall\nliblocale-gettext-perl\t\t\t\tinstall\nliblockfile-bin\t\t\t\t\tinstall\nliblockfile1\t\t\t\t\tinstall\nliblwres80\t\t\t\t\tinstall\nliblzma5\t\t\t\t\tinstall\nlibmagic1\t\t\t\t\tinstall\nlibmount1\t\t\t\t\tinstall\nlibncurses5\t\t\t\t\tinstall\nlibncursesw5\t\t\t\t\tinstall\nlibnewt0.52\t\t\t\t\tinstall\nlibnfnetlink0\t\t\t\t\tinstall\nlibnih-dbus1\t\t\t\t\tinstall\nlibnih1\t\t\t\t\t\tinstall\nlibnl-3-200\t\t\t\t\tinstall\nlibnl-genl-3-200\t\t\t\tinstall\nlibp11-kit0\t\t\t\t\tinstall\nlibpam-modules\t\t\t\t\tinstall\nlibpam-modules-bin\t\t\t\tinstall\nlibpam-runtime\t\t\t\t\tinstall\nlibpam0g\t\t\t\t\tinstall\nlibparse-debianchangelog-perl\t\t\tinstall\nlibparted0debian1\t\t\t\tinstall\nlibpcap0.8\t\t\t\t\tinstall\nlibpci3\t\t\t\t\t\tinstall\nlibpciaccess0\t\t\t\t\tinstall\nlibpcre3\t\t\t\t\tinstall\nlibpcsclite1\t\t\t\t\tinstall\nlibpipeline1\t\t\t\t\tinstall\nlibplymouth2\t\t\t\t\tinstall\nlibpng12-0\t\t\t\t\tinstall\nlibpolkit-gobject-1-0\t\t\t\tinstall\nlibpopt0\t\t\t\t\tinstall\nlibpython2.7\t\t\t\t\tinstall\nlibreadline6\t\t\t\t\tinstall\nlibroken18-heimdal\t\t\t\tinstall\nlibrtmp0\t\t\t\t\tinstall\nlibsasl2-2\t\t\t\t\tinstall\nlibsasl2-modules\t\t\t\tinstall\nlibselinux1\t\t\t\t\tinstall\nlibsigc++-2.0-0c2a\t\t\t\tinstall\nlibslang2\t\t\t\t\tinstall\nlibsqlite3-0\t\t\t\t\tinstall\nlibss2\t\t\t\t\t\tinstall\nlibssl1.0.0\t\t\t\t\tinstall\nlibstdc++6\t\t\t\t\tinstall\nlibsub-name-perl\t\t\t\tinstall\nlibswitch-perl\t\t\t\t\tinstall\nlibtasn1-3\t\t\t\t\tinstall\nlibtext-charwidth-perl\t\t\t\tinstall\nlibtext-iconv-perl\t\t\t\tinstall\nlibtext-wrapi18n-perl\t\t\t\tinstall\nlibtimedate-perl\t\t\t\tinstall\nlibtinfo5\t\t\t\t\tinstall\nlibudev0\t\t\t\t\tinstall\nlibusb-0.1-4\t\t\t\t\tinstall\nlibusb-1.0-0\t\t\t\t\tinstall\nlibuuid1\t\t\t\t\tinstall\nlibwind0-heimdal\t\t\t\tinstall\nlibwrap0\t\t\t\t\tinstall\nlibx11-6\t\t\t\t\tinstall\nlibx11-data\t\t\t\t\tinstall\nlibxapian22\t\t\t\t\tinstall\nlibxau6\t\t\t\t\t\tinstall\nlibxcb1\t\t\t\t\t\tinstall\nlibxdmcp6\t\t\t\t\tinstall\nlibxext6\t\t\t\t\tinstall\nlibxml2\t\t\t\t\t\tinstall\nlibxmuu1\t\t\t\t\tinstall\nlibyaml-0-2\t\t\t\t\tinstall\nlinux-firmware\t\t\t\t\tinstall\nlinux-headers-3.2.0-60\t\t\t\tinstall\nlinux-headers-3.2.0-60-virtual\t\t\tinstall\nlinux-headers-virtual\t\t\t\tinstall\nlinux-image-3.2.0-60-virtual\t\t\tinstall\nlinux-image-virtual\t\t\t\tinstall\nlinux-virtual\t\t\t\t\tinstall\nlocales\t\t\t\t\t\tinstall\nlockfile-progs\t\t\t\t\tinstall\nlogin\t\t\t\t\t\tinstall\nlogrotate\t\t\t\t\tinstall\nlsb-base\t\t\t\t\tinstall\nlsb-release\t\t\t\t\tinstall\nlshw\t\t\t\t\t\tinstall\nlsof\t\t\t\t\t\tinstall\nltrace\t\t\t\t\t\tinstall\nmakedev\t\t\t\t\t\tinstall\nman-db\t\t\t\t\t\tinstall\nmanpages\t\t\t\t\tinstall\nmawk\t\t\t\t\t\tinstall\nmemtest86+\t\t\t\t\tinstall\nmime-support\t\t\t\t\tinstall\nmlocate\t\t\t\t\t\tinstall\nmodule-init-tools\t\t\t\tinstall\nmount\t\t\t\t\t\tinstall\nmountall\t\t\t\t\tinstall\nmtr-tiny\t\t\t\t\tinstall\nmultiarch-support\t\t\t\tinstall\nnano\t\t\t\t\t\tinstall\nncurses-base\t\t\t\t\tinstall\nncurses-bin\t\t\t\t\tinstall\nnet-tools\t\t\t\t\tinstall\nnetbase\t\t\t\t\t\tinstall\nnetcat-openbsd\t\t\t\t\tinstall\nntfs-3g\t\t\t\t\t\tinstall\nntpdate\t\t\t\t\t\tinstall\nopenssh-client\t\t\t\t\tinstall\nopenssh-server\t\t\t\t\tinstall\nopenssl\t\t\t\t\t\tinstall\nos-prober\t\t\t\t\tinstall\nparted\t\t\t\t\t\tinstall\npasswd\t\t\t\t\t\tinstall\npatch\t\t\t\t\t\tinstall\npciutils\t\t\t\t\tinstall\nperl\t\t\t\t\t\tinstall\nperl-base\t\t\t\t\tinstall\nperl-modules\t\t\t\t\tinstall\nplymouth\t\t\t\t\tinstall\nplymouth-theme-ubuntu-text\t\t\tinstall\npopularity-contest\t\t\t\tinstall\npowermgmt-base\t\t\t\t\tinstall\nppp\t\t\t\t\t\tinstall\npppconfig\t\t\t\t\tinstall\npppoeconf\t\t\t\t\tinstall\nprocps\t\t\t\t\t\tinstall\npsmisc\t\t\t\t\t\tinstall\npython\t\t\t\t\t\tinstall\npython-apport\t\t\t\t\tinstall\npython-apt\t\t\t\t\tinstall\npython-apt-common\t\t\t\tinstall\npython-boto\t\t\t\t\tinstall\npython-chardet\t\t\t\t\tinstall\npython-cheetah\t\t\t\t\tinstall\npython-configobj\t\t\t\tinstall\npython-crypto\t\t\t\t\tinstall\npython-dbus\t\t\t\t\tinstall\npython-dbus-dev\t\t\t\t\tinstall\npython-debian\t\t\t\t\tinstall\npython-gdbm\t\t\t\t\tinstall\npython-gi\t\t\t\t\tinstall\npython-gnupginterface\t\t\t\tinstall\npython-httplib2\t\t\t\t\tinstall\npython-keyring\t\t\t\t\tinstall\npython-launchpadlib\t\t\t\tinstall\npython-lazr.restfulclient\t\t\tinstall\npython-lazr.uri\t\t\t\t\tinstall\npython-m2crypto\t\t\t\t\tinstall\npython-minimal\t\t\t\t\tinstall\npython-newt\t\t\t\t\tinstall\npython-oauth\t\t\t\t\tinstall\npython-openssl\t\t\t\t\tinstall\npython-pam\t\t\t\t\tinstall\npython-paramiko\t\t\t\t\tinstall\npython-pkg-resources\t\t\t\tinstall\npython-problem-report\t\t\t\tinstall\npython-pycurl\t\t\t\t\tinstall\npython-serial\t\t\t\t\tinstall\npython-simplejson\t\t\t\tinstall\npython-software-properties\t\t\tinstall\npython-twisted-bin\t\t\t\tinstall\npython-twisted-core\t\t\t\tinstall\npython-twisted-names\t\t\t\tinstall\npython-twisted-web\t\t\t\tinstall\npython-wadllib\t\t\t\t\tinstall\npython-xapian\t\t\t\t\tinstall\npython-yaml\t\t\t\t\tinstall\npython-zope.interface\t\t\t\tinstall\npython2.7\t\t\t\t\tinstall\npython2.7-minimal\t\t\t\tinstall\nreadline-common\t\t\t\t\tinstall\nresolvconf\t\t\t\t\tinstall\nrsync\t\t\t\t\t\tinstall\nrsyslog\t\t\t\t\t\tinstall\nscreen\t\t\t\t\t\tinstall\nsed\t\t\t\t\t\tinstall\nsensible-utils\t\t\t\t\tinstall\nsgml-base\t\t\t\t\tinstall\nssh-import-id\t\t\t\t\tinstall\nstrace\t\t\t\t\t\tinstall\nsudo\t\t\t\t\t\tinstall\nsysv-rc\t\t\t\t\t\tinstall\nsysvinit-utils\t\t\t\t\tinstall\ntar\t\t\t\t\t\tinstall\ntasksel\t\t\t\t\t\tinstall\ntasksel-data\t\t\t\t\tinstall\ntcpd\t\t\t\t\t\tinstall\ntcpdump\t\t\t\t\t\tinstall\ntelnet\t\t\t\t\t\tinstall\ntime\t\t\t\t\t\tinstall\ntmux\t\t\t\t\t\tinstall\ntzdata\t\t\t\t\t\tinstall\nubuntu-keyring\t\t\t\t\tinstall\nubuntu-minimal\t\t\t\t\tinstall\nubuntu-standard\t\t\t\t\tinstall\nucf\t\t\t\t\t\tinstall\nudev\t\t\t\t\t\tinstall\nufw\t\t\t\t\t\tinstall\nunattended-upgrades\t\t\t\tinstall\nupdate-manager-core\t\t\t\tinstall\nupdate-notifier-common\t\t\t\tinstall\nupstart\t\t\t\t\t\tinstall\nureadahead\t\t\t\t\tinstall\nusbutils\t\t\t\t\tinstall\nutil-linux\t\t\t\t\tinstall\nuuid-runtime\t\t\t\t\tinstall\nvim\t\t\t\t\t\tinstall\nvim-common\t\t\t\t\tinstall\nvim-runtime\t\t\t\t\tinstall\nvim-tiny\t\t\t\t\tinstall\nw3m\t\t\t\t\t\tinstall\nwget\t\t\t\t\t\tinstall\nwhiptail\t\t\t\t\tinstall\nwhoopsie\t\t\t\t\tinstall\nwireless-regdb\t\t\t\t\tinstall\nwireless-tools\t\t\t\t\tinstall\nwpasupplicant\t\t\t\t\tinstall\nxauth\t\t\t\t\t\tinstall\nxkb-data\t\t\t\t\tinstall\nxml-core\t\t\t\t\tinstall\nxz-lzma\t\t\t\t\t\tinstall\nxz-utils\t\t\t\t\tinstall\nzlib1g\t\t\t\t\t\tinstall\n#> [automated-admin-user: install]: Packages : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho '[automated-admin-user: install]: Packages...';\n{\n{ debconf-set-selections <<EOF\ndebconf debconf/frontend select noninteractive\ndebconf debconf/frontend seen false\nEOF\n} && enableStart() {\nrm /usr/sbin/policy-rc.d\n} && apt-get -q -y install sudo+ && dpkg --get-selections\n } || { echo '#> [automated-admin-user: install]: Packages : FAIL'; exit 1;} >&2 \necho '#> [automated-admin-user: install]: Packages : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package, | |
:context "[automated-admin-user: install]: "} | |
{:exit 0, | |
:out "", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent passwd tom; then /usr/sbin/usermod --shell \"/bin/bash\" tom;else /usr/sbin/useradd --shell \"/bin/bash\" --create-home tom;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/user, | |
:context "[automated-admin-user]: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/...\n#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/...';\n{\nmkdir -m \"755\" -p $(getent passwd tom | cut -d: -f6)/.ssh/ && chown --recursive tom $(getent passwd tom | cut -d: -f6)/.ssh/ && chmod 755 $(getent passwd tom | cut -d: -f6)/.ssh/\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: Directory $(getent passwd tom | cut -d: -f6)/.ssh/ : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys...\n#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys...';\n{\ntouch $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && chown tom $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && chmod 644 $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: file $(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/file, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32)...\n#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32)...';\n{\nauth_file=$(getent passwd tom | cut -d: -f6)/.ssh/authorized_keys && if ! ( fgrep \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDnV92GSukTdPwSitdUzox9hMlwYz3oApmJ9FY+7GQ/QNMfLD7jFDmn9wLqJU6WpWtXd5S+RfILZzS0LvPgFgWrY8/ZxcwwVlnp0adGdxb3s3iwsnSLFwF2KhyI6rpbDuFZaO+0ExbNUWC8YBXcp/9wQlsZ2OUJ/jQkpqwZ+L0sLsGS2cSH1lGwpYUr/oZ15gbSZ6mvqzBbZnZXI3IABd23b7xdaAldV3PYHCdJ6zSUNa675YsTj+SdEBPQpx0IakcH+XAp62ePchxMocEKeeTfW/aYDscxBw7M9aGZTCsNl4LKMzeF7FYlPAF1VYtXg31YxACKEpRHW+IVY9oSDenV tom@infolace.com\" ${auth_file} ); then\necho \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDnV92GSukTdPwSitdUzox9hMlwYz3oApmJ9FY+7GQ/QNMfLD7jFDmn9wLqJU6WpWtXd5S+RfILZzS0LvPgFgWrY8/ZxcwwVlnp0adGdxb3s3iwsnSLFwF2KhyI6rpbDuFZaO+0ExbNUWC8YBXcp/9wQlsZ2OUJ/jQkpqwZ+L0sLsGS2cSH1lGwpYUr/oZ15gbSZ6mvqzBbZnZXI3IABd23b7xdaAldV3PYHCdJ6zSUNa675YsTj+SdEBPQpx0IakcH+XAp62ePchxMocEKeeTfW/aYDscxBw7M9aGZTCsNl4LKMzeF7FYlPAF1VYtXg31YxACKEpRHW+IVY9oSDenV tom@infolace.com\n\" >> ${auth_file}\nfi\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: authorize-key on user tom (ssh_key.clj:32) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37)...\n#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37)...';\n{\nif hash chcon 2>&- && [ -d /etc/selinux ] && [ -e /selinux/enforce ] && stat --format %C $(getent passwd tom | cut -d: -f6)/.ssh/ 2>&-; then chcon -Rv --type=user_home_t $(getent passwd tom | cut -d: -f6)/.ssh/;fi\n } || { echo '#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: authorize-user-key: authorize-key: Set selinux permissions (ssh_key.clj:37) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"automated-admin-user: authorize-user-key: authorize-key: "} | |
{:exit 0, | |
:out | |
"automated-admin-user: sudoers: remote-file /etc/sudoers...\n--- /etc/sudoers\t2012-01-31 15:56:42.000000000 +0000\n+++ /tmp/root/bBCdbpTJf7UZAHwfBogFgg/sudoers\t2014-04-24 16:56:04.024528263 +0000\n@@ -1,29 +1,4 @@\n-#\n-# This file MUST be edited with the 'visudo' command as root.\n-#\n-# Please consider adding local content in /etc/sudoers.d/ instead of\n-# directly modifying this file.\n-#\n-# See the man page for details on how to write a sudoers file.\n-#\n-Defaults\tenv_reset\n-Defaults\tsecure_path=\"/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin\"\n-\n-# Host alias specification\n-\n-# User alias specification\n-\n-# Cmnd alias specification\n-\n-# User privilege specification\n-root\tALL=(ALL:ALL) ALL\n-\n-# Members of the admin group may gain root privileges\n-%admin ALL=(ALL) ALL\n-\n-# Allow members of group sudo to execute any command\n-%sudo\tALL=(ALL:ALL) ALL\n-\n-# See sudoers(5) for more information on \"#include\" directives:\n-\n-#includedir /etc/sudoers.d\n+Defaults env_keep=SSH_AUTH_SOCK\n+root ALL = (ALL) ALL\n+%adm ALL = (ALL) ALL\n+tom ALL = (ALL) NOPASSWD: ALL\nMD5 sum is 46780992b159bec006d30e7c37b65c1c sudoers\n#> automated-admin-user: sudoers: remote-file /etc/sudoers : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'automated-admin-user: sudoers: remote-file /etc/sudoers...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /etc/sudoers /var/lib/pallet/etc/sudoers /var/lib/pallet/etc/sudoers.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) <<EOFpallet\nDefaults env_keep=SSH_AUTH_SOCK\nroot ALL = (ALL) ALL\n%adm ALL = (ALL) ALL\ntom ALL = (ALL) NOPASSWD: ALL\nEOFpallet\n } && contentdiff= && if [ -e /etc/sudoers ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) ]; then\ndiff -u /etc/sudoers ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && chgrp $(id -ng root) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && chmod 0440 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/bBCdbpTJf7UZAHwfBogFgg/$(basename /etc/sudoers) /etc/sudoers && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /etc/sudoers /var/lib/pallet/etc/sudoers\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /etc/sudoers /var/lib/pallet/etc/sudoers\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /etc/sudoers /var/lib/pallet/etc/sudoers.md5\n}\nfi\n } || { echo '#> automated-admin-user: sudoers: remote-file /etc/sudoers : FAIL'; exit 1;} >&2 \necho '#> automated-admin-user: sudoers: remote-file /etc/sudoers : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "automated-admin-user: sudoers: "}), | |
:phase :bootstrap} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"[install: install: install-from]: Packages...\nReading package lists...\nBuilding dependency tree...\nReading state information...\nThe following extra packages will be installed:\n ca-certificates-java consolekit cpp cpp-4.6 cryptsetup-bin dbus-x11\n dconf-gsettings-backend dconf-service fontconfig fontconfig-config\n gconf-service gconf-service-backend gconf2 gconf2-common gvfs gvfs-common\n gvfs-daemons gvfs-libs hicolor-icon-theme icedtea-7-jre-jamvm java-common\n libasound2 libasyncns0 libatasmart4 libatk-wrapper-java\n libatk-wrapper-java-jni libatk1.0-0 libatk1.0-data libavahi-client3\n libavahi-common-data libavahi-common3 libavahi-glib1 libbonobo2-0\n libbonobo2-common libcairo-gobject2 libcairo2 libcanberra0 libck-connector0\n libcryptsetup4 libcups2 libdatrie1 libdconf0 libdevmapper-event1.02.1\n libflac8 libfontconfig1 libfontenc1 libgconf-2-4 libgconf2-4\n libgdk-pixbuf2.0-0 libgdk-pixbuf2.0-common libgdu0 libgif4 libgl1-mesa-dri\n libgl1-mesa-glx libglapi-mesa libgnome-keyring-common libgnome-keyring0\n libgnome2-0 libgnome2-bin libgnome2-common libgnomevfs2-0\n libgnomevfs2-common libgtk-3-0 libgtk-3-bin libgtk-3-common libgtk2.0-0\n libgtk2.0-bin libgtk2.0-common libgudev-1.0-0 libice-dev libice6\n libidl-common libidl0 libjasper1 libjpeg-turbo8 libjpeg8 libjson0 liblcms2-2\n libllvm3.0 libltdl7 liblvm2app2.2 libmpc2 libmpfr4 libnspr4 libnss3\n libnss3-1d libogg0 liborbit2 libpam-ck-connector libpango1.0-0 libpixman-1-0\n libpolkit-agent-1-0 libpolkit-backend-1-0 libpthread-stubs0\n libpthread-stubs0-dev libpulse0 libsgutils2-2 libsm-dev libsm6 libsndfile1\n libtdb1 libthai-data libthai0 libtiff4 libvorbis0a libvorbisenc2\n libvorbisfile3 libx11-dev libx11-doc libx11-xcb1 libxau-dev libxaw7\n libxcb-glx0 libxcb-render0 libxcb-shape0 libxcb-shm0 libxcb1-dev\n libxcomposite1 libxcursor1 libxdamage1 libxdmcp-dev libxfixes3 libxft2\n libxi6 libxinerama1 libxmu6 libxpm4 libxrandr2 libxrender1 libxt-dev libxt6\n libxtst6 libxv1 libxxf86dga1 libxxf86vm1 mtools openjdk-7-jre\n openjdk-7-jre-headless policykit-1 policykit-1-gnome shared-mime-info\n sound-theme-freedesktop ttf-dejavu-core ttf-dejavu-extra tzdata-java udisks\n x11-common x11-utils x11proto-core-dev x11proto-input-dev x11proto-kb-dev\n xorg-sgml-doctools xtrans-dev\nSuggested packages:\n cpp-doc gcc-4.6-locales gconf-defaults-service gvfs-backends default-jre\n equivs libasound2-plugins libasound2-python libbonobo2-bin libcanberra-gtk0\n libcanberra-pulse cups-common libglide3 gnome-keyring desktop-base\n libgnomevfs2-bin libgnomevfs2-extra gamin fam gnome-mime-data\n librsvg2-common libjasper-runtime liblcms2-utils ttf-baekmuk\n ttf-arphic-gbsn00lp ttf-arphic-bsmi00lp ttf-arphic-gkai00mp\n ttf-arphic-bkai00mp pulseaudio sg3-utils libxcb-doc floppyd openjdk-7-demo\n openjdk-7-source visualvm icedtea-7-plugin libnss-mdns sun-java6-fonts\n fonts-ipafont-gothic fonts-ipafont-mincho ttf-wqy-microhei ttf-wqy-zenhei\n ttf-indic-fonts-core ttf-telugu-fonts ttf-oriya-fonts ttf-kannada-fonts\n ttf-bengali-fonts xfsprogs reiserfsprogs mdadm mesa-utils\nThe following NEW packages will be installed:\n ca-certificates-java consolekit cpp cpp-4.6 cryptsetup-bin dbus-x11\n dconf-gsettings-backend dconf-service fontconfig fontconfig-config\n gconf-service gconf-service-backend gconf2 gconf2-common gvfs gvfs-common\n gvfs-daemons gvfs-libs hicolor-icon-theme icedtea-7-jre-jamvm java-common\n libasound2 libasyncns0 libatasmart4 libatk-wrapper-java\n libatk-wrapper-java-jni libatk1.0-0 libatk1.0-data libavahi-client3\n libavahi-common-data libavahi-common3 libavahi-glib1 libbonobo2-0\n libbonobo2-common libcairo-gobject2 libcairo2 libcanberra0 libck-connector0\n libcryptsetup4 libcups2 libdatrie1 libdconf0 libdevmapper-event1.02.1\n libflac8 libfontconfig1 libfontenc1 libgconf-2-4 libgconf2-4\n libgdk-pixbuf2.0-0 libgdk-pixbuf2.0-common libgdu0 libgif4 libgl1-mesa-dri\n libgl1-mesa-glx libglapi-mesa libgnome-keyring-common libgnome-keyring0\n libgnome2-0 libgnome2-bin libgnome2-common libgnomevfs2-0\n libgnomevfs2-common libgtk-3-0 libgtk-3-bin libgtk-3-common libgtk2.0-0\n libgtk2.0-bin libgtk2.0-common libgudev-1.0-0 libice-dev libice6\n libidl-common libidl0 libjasper1 libjpeg-turbo8 libjpeg8 libjson0 liblcms2-2\n libllvm3.0 libltdl7 liblvm2app2.2 libmpc2 libmpfr4 libnspr4 libnss3\n libnss3-1d libogg0 liborbit2 libpam-ck-connector libpango1.0-0 libpixman-1-0\n libpolkit-agent-1-0 libpolkit-backend-1-0 libpthread-stubs0\n libpthread-stubs0-dev libpulse0 libsgutils2-2 libsm-dev libsm6 libsndfile1\n libtdb1 libthai-data libthai0 libtiff4 libvorbis0a libvorbisenc2\n libvorbisfile3 libx11-dev libx11-doc libx11-xcb1 libxau-dev libxaw7\n libxcb-glx0 libxcb-render0 libxcb-shape0 libxcb-shm0 libxcb1-dev\n libxcomposite1 libxcursor1 libxdamage1 libxdmcp-dev libxfixes3 libxft2\n libxi6 libxinerama1 libxmu6 libxpm4 libxrandr2 libxrender1 libxt-dev libxt6\n libxtst6 libxv1 libxxf86dga1 libxxf86vm1 mtools openjdk-7-jdk openjdk-7-jre\n openjdk-7-jre-headless policykit-1 policykit-1-gnome shared-mime-info\n sound-theme-freedesktop ttf-dejavu-core ttf-dejavu-extra tzdata-java udisks\n x11-common x11-utils x11proto-core-dev x11proto-input-dev x11proto-kb-dev\n xorg-sgml-doctools xtrans-dev\n0 upgraded, 154 newly installed, 0 to remove and 8 not upgraded.\nNeed to get 100 MB of archives.\nAfter this operation, 220 MB of additional disk space will be used.\nGet:1 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libasound2 amd64 1.0.25-1ubuntu10.2 [428 kB]\nGet:2 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libasyncns0 amd64 0.8-4 [13.2 kB]\nGet:3 http://security.ubuntu.com/ubuntu/ precise-security/main libcups2 amd64 1.5.3-0ubuntu8.2 [171 kB]\nGet:4 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatasmart4 amd64 0.18-3 [26.9 kB]\nGet:5 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk1.0-data all 2.4.0-0ubuntu1 [12.7 kB]\nGet:6 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk1.0-0 amd64 2.4.0-0ubuntu1 [60.1 kB]\nGet:7 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-common all 2.24.10-0ubuntu6.1 [118 kB]\nGet:8 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ttf-dejavu-core all 2.33-2ubuntu1 [1,552 kB]\nGet:9 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main fontconfig-config all 2.8.0-3ubuntu9.1 [44.4 kB]\nGet:10 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libfontconfig1 amd64 2.8.0-3ubuntu9.1 [125 kB]\nGet:11 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpixman-1-0 amd64 0.30.2-1ubuntu0.0.0.0.1 [256 kB]\nGet:12 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-render0 amd64 1.8.1-1ubuntu0.2 [13.3 kB]\nGet:13 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-shm0 amd64 1.8.1-1ubuntu0.2 [5,516 B]\nGet:14 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxrender1 amd64 1:0.9.6-2ubuntu0.1 [20.6 kB]\nGet:15 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libcairo2 amd64 1.10.2-6.1ubuntu3 [467 kB]\nGet:16 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-common-data amd64 0.6.30-5ubuntu2.1 [21.4 kB]\nGet:17 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-common3 amd64 0.6.30-5ubuntu2.1 [25.5 kB]\nGet:18 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-client3 amd64 0.6.30-5ubuntu2.1 [30.3 kB]\nGet:19 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libjpeg-turbo8 amd64 1.1.90+svn733-0ubuntu4.3 [111 kB]\nGet:20 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjpeg8 amd64 8c-2ubuntu7 [2,112 B]\nGet:21 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjasper1 amd64 1.900.1-13 [154 kB]\nGet:22 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libtiff4 amd64 3.9.5-2ubuntu1.5 [145 kB]\nGet:23 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgdk-pixbuf2.0-common all 2.26.1-1ubuntu1.1 [7,846 B]\nGet:24 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgdk-pixbuf2.0-0 amd64 2.26.1-1ubuntu1.1 [200 kB]\nGet:25 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libthai-data all 0.1.16-3 [179 kB]\nGet:26 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libdatrie1 amd64 0.2.5-3 [16.4 kB]\nGet:27 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libthai0 amd64 0.1.16-3 [19.7 kB]\nGet:28 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxft2 amd64 2.2.0-3ubuntu2 [42.3 kB]\nGet:29 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main fontconfig amd64 2.8.0-3ubuntu9.1 [156 kB]\nGet:30 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpango1.0-0 amd64 1.30.0-0ubuntu3.1 [363 kB]\nGet:31 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxcomposite1 amd64 1:0.4.3-2build1 [7,708 B]\nGet:32 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxfixes3 amd64 1:5.0-4ubuntu4.2 [12.5 kB]\nGet:33 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcursor1 amd64 1:1.1.12-1ubuntu0.1 [22.7 kB]\nGet:34 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxdamage1 amd64 1:1.1.3-2build1 [7,560 B]\nGet:35 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxi6 amd64 2:1.7.1.901-1ubuntu1~precise1 [31.6 kB]\nGet:36 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxinerama1 amd64 2:1.1.1-3ubuntu0.1 [8,072 B]\nGet:37 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxrandr2 amd64 2:1.3.2-2ubuntu0.2 [17.5 kB]\nGet:38 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main shared-mime-info amd64 1.0-0ubuntu4.1 [496 kB]\nGet:39 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-0 amd64 2.24.10-0ubuntu6.1 [2,674 kB]\nGet:40 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnspr4 amd64 4.9.5-0ubuntu0.12.04.2 [141 kB]\nGet:41 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnss3 amd64 3.15.4-0ubuntu0.12.04.2 [1,229 kB]\nGet:42 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnss3-1d amd64 3.15.4-0ubuntu0.12.04.2 [13.4 kB]\nGet:43 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ca-certificates-java all 20110912ubuntu6 [8,186 B]\nGet:44 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main tzdata-java all 2014a-0ubuntu0.12.04 [126 kB]\nGet:45 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main java-common all 0.43ubuntu2 [61.7 kB]\nGet:46 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liblcms2-2 amd64 2.2+git20110628-2ubuntu3.1 [143 kB]\nGet:47 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jre-headless amd64 7u51-2.4.4-0ubuntu0.12.04.2 [41.4 MB]\nGet:48 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgif4 amd64 4.1.6-9ubuntu1 [31.0 kB]\nGet:49 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjson0 amd64 0.9-1ubuntu1 [16.3 kB]\nGet:50 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libogg0 amd64 1.2.2~dfsg-1ubuntu1 [17.2 kB]\nGet:51 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libflac8 amd64 1.2.1-6 [144 kB]\nGet:52 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbis0a amd64 1.3.2-1ubuntu3 [103 kB]\nGet:53 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbisenc2 amd64 1.3.2-1ubuntu3 [136 kB]\nGet:54 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsndfile1 amd64 1.0.25-4 [163 kB]\nGet:55 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpulse0 amd64 1:1.1-0ubuntu15.4 [289 kB]\nGet:56 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11-common all 1:7.6+12ubuntu2 [52.0 kB]\nGet:57 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxtst6 amd64 2:1.2.0-4ubuntu0.1 [13.9 kB]\nGet:58 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libglapi-mesa amd64 8.0.4-0ubuntu0.7 [21.3 kB]\nGet:59 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-xcb1 amd64 2:1.4.99.1-0ubuntu2.2 [10.5 kB]\nGet:60 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-glx0 amd64 1.8.1-1ubuntu0.2 [26.4 kB]\nGet:61 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxxf86vm1 amd64 1:1.1.1-2ubuntu0.1 [11.9 kB]\nGet:62 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgl1-mesa-glx amd64 8.0.4-0ubuntu0.7 [104 kB]\nGet:63 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jre amd64 7u51-2.4.4-0ubuntu0.12.04.2 [226 kB]\nGet:64 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk-wrapper-java all 0.30.4-0ubuntu2 [30.9 kB]\nGet:65 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk-wrapper-java-jni amd64 0.30.4-0ubuntu2 [30.5 kB]\nGet:66 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-glib1 amd64 0.6.30-5ubuntu2.1 [8,838 B]\nGet:67 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libbonobo2-common all 2.32.1-0ubuntu1.1 [37.4 kB]\nGet:68 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libmpfr4 amd64 3.1.0-3ubuntu2 [201 kB]\nGet:69 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libmpc2 amd64 0.9-4 [39.5 kB]\nGet:70 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cpp-4.6 amd64 4.6.3-1ubuntu5 [4,836 kB]\nGet:71 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cpp amd64 4:4.6.3-1ubuntu5 [27.6 kB]\nGet:72 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libidl-common all 0.8.14-0.2ubuntu2 [8,572 B]\nGet:73 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libidl0 amd64 0.8.14-0.2ubuntu2 [81.6 kB]\nGet:74 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liborbit2 amd64 1:2.14.19-0.1ubuntu1 [186 kB]\nGet:75 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libbonobo2-0 amd64 2.32.1-0ubuntu1.1 [285 kB]\nGet:76 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libcairo-gobject2 amd64 1.10.2-6.1ubuntu3 [20.5 kB]\nGet:77 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libltdl7 amd64 2.4.2-1ubuntu1 [37.6 kB]\nGet:78 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libtdb1 amd64 1.2.9-4 [35.3 kB]\nGet:79 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbisfile3 amd64 1.3.2-1ubuntu3 [17.7 kB]\nGet:80 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main sound-theme-freedesktop all 0.7.pristine-2 [412 kB]\nGet:81 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libcanberra0 amd64 0.28-3ubuntu3 [43.8 kB]\nGet:82 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf2-common all 3.2.5-0ubuntu2 [22.1 kB]\nGet:83 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgconf-2-4 amd64 3.2.5-0ubuntu2 [97.2 kB]\nGet:84 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgnome-keyring-common all 3.2.2-2 [5,060 B]\nGet:85 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgnome-keyring0 amd64 3.2.2-2 [62.6 kB]\nGet:86 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgudev-1.0-0 amd64 1:175-0ubuntu9.5 [14.5 kB]\nGet:87 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libdevmapper-event1.02.1 amd64 2:1.02.48-4ubuntu7.4 [10.4 kB]\nGet:88 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liblvm2app2.2 amd64 2.02.66-4ubuntu7.4 [223 kB]\nGet:89 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsgutils2-2 amd64 1.33-1 [64.4 kB]\nGet:90 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main udisks amd64 1.0.4-5ubuntu2.2 [250 kB]\nGet:91 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgdu0 amd64 3.0.2-2ubuntu7 [90.0 kB]\nGet:92 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf-service-backend amd64 3.2.5-0ubuntu2 [74.1 kB]\nGet:93 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf-service amd64 3.2.5-0ubuntu2 [2,012 B]\nGet:94 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dbus-x11 amd64 1.4.18-1ubuntu1.4 [22.1 kB]\nGet:95 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf2 amd64 3.2.5-0ubuntu2 [79.7 kB]\nGet:96 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnomevfs2-common amd64 1:2.24.4-1ubuntu2.1 [24.6 kB]\nGet:97 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnomevfs2-0 amd64 1:2.24.4-1ubuntu2.1 [275 kB]\nGet:98 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-common all 2.32.1-2ubuntu1.1 [36.9 kB]\nGet:99 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-bin amd64 2.32.1-2ubuntu1.1 [15.9 kB]\nGet:100 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-0 amd64 2.32.1-2ubuntu1.1 [53.4 kB]\nGet:101 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libdconf0 amd64 0.12.0-0ubuntu1.1 [25.4 kB]\nGet:102 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dconf-service amd64 0.12.0-0ubuntu1.1 [16.9 kB]\nGet:103 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dconf-gsettings-backend amd64 0.12.0-0ubuntu1.1 [16.1 kB]\nGet:104 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-common all 3.4.2-0ubuntu0.8 [145 kB]\nGet:105 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-0 amd64 3.4.2-0ubuntu0.8 [2,281 kB]\nGet:106 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libice6 amd64 2:1.0.7-2build1 [46.1 kB]\nGet:107 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libllvm3.0 amd64 3.0-4ubuntu1 [7,456 kB]\nGet:108 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpolkit-agent-1-0 amd64 0.104-1ubuntu1.1 [17.7 kB]\nGet:109 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpolkit-backend-1-0 amd64 0.104-1ubuntu1.1 [41.7 kB]\nGet:110 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsm6 amd64 2:1.2.0-2build1 [18.1 kB]\nGet:111 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxt6 amd64 1:1.1.1-2ubuntu0.1 [186 kB]\nGet:112 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxmu6 amd64 2:1.1.0-3 [52.7 kB]\nGet:113 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxpm4 amd64 1:3.5.9-4 [37.0 kB]\nGet:114 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxaw7 amd64 2:1.0.9-3ubuntu1 [201 kB]\nGet:115 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-shape0 amd64 1.8.1-1ubuntu0.2 [6,244 B]\nGet:116 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxv1 amd64 2:1.0.6-2ubuntu0.1 [12.4 kB]\nGet:117 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libck-connector0 amd64 0.4.5-2ubuntu0.1 [10.4 kB]\nGet:118 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main consolekit amd64 0.4.5-2ubuntu0.1 [120 kB]\nGet:119 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libcryptsetup4 amd64 2:1.4.1-2ubuntu4 [55.8 kB]\nGet:120 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cryptsetup-bin amd64 2:1.4.1-2ubuntu4 [32.2 kB]\nGet:121 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libfontenc1 amd64 1:1.1.0-1 [15.4 kB]\nGet:122 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxxf86dga1 amd64 2:1.1.2-1ubuntu0.1 [13.9 kB]\nGet:123 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11-utils amd64 7.6+4ubuntu0.1 [230 kB]\nGet:124 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-common all 1.12.1-0ubuntu1.2 [16.4 kB]\nGet:125 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-libs amd64 1.12.1-0ubuntu1.2 [52.8 kB]\nGet:126 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-daemons amd64 1.12.1-0ubuntu1.2 [348 kB]\nGet:127 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs amd64 1.12.1-0ubuntu1.2 [98.9 kB]\nGet:128 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main hicolor-icon-theme all 0.12-1ubuntu2 [10.2 kB]\nGet:129 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgconf2-4 amd64 3.2.5-0ubuntu2 [2,016 B]\nGet:130 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgl1-mesa-dri amd64 8.0.4-0ubuntu0.7 [3,035 kB]\nGet:131 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-bin amd64 3.4.2-0ubuntu0.8 [15.9 kB]\nGet:132 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-bin amd64 2.24.10-0ubuntu6.1 [10.3 kB]\nGet:133 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main xorg-sgml-doctools all 1:1.10-1 [12.0 kB]\nGet:134 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11proto-core-dev all 7.0.22-1ubuntu0.1 [300 kB]\nGet:135 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libice-dev amd64 2:1.0.7-2build1 [133 kB]\nGet:136 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpam-ck-connector amd64 0.4.5-2ubuntu0.1 [8,188 B]\nGet:137 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libpthread-stubs0 amd64 0.3-3 [3,258 B]\nGet:138 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libpthread-stubs0-dev amd64 0.3-3 [2,866 B]\nGet:139 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsm-dev amd64 2:1.2.0-2build1 [89.5 kB]\nGet:140 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxau-dev amd64 1:1.0.6-4 [10.5 kB]\nGet:141 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxdmcp-dev amd64 1:1.1.0-4 [26.9 kB]\nGet:142 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11proto-input-dev all 2.3-1~precise1 [139 kB]\nGet:143 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main x11proto-kb-dev all 1.0.5-2 [27.6 kB]\nGet:144 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main xtrans-dev all 1.2.6-2 [82.9 kB]\nGet:145 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb1-dev amd64 1.8.1-1ubuntu0.2 [82.5 kB]\nGet:146 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-dev amd64 2:1.4.99.1-0ubuntu2.2 [912 kB]\nGet:147 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-doc all 2:1.4.99.1-0ubuntu2.2 [2,435 kB]\nGet:148 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxt-dev amd64 1:1.1.1-2ubuntu0.1 [490 kB]\nGet:149 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main mtools amd64 4.0.12-1ubuntu0.12.04.1 [186 kB]\nGet:150 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jdk amd64 7u51-2.4.4-0ubuntu0.12.04.2 [16.4 MB]\nGet:151 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main policykit-1 amd64 0.104-1ubuntu1.1 [54.9 kB]\nGet:152 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main policykit-1-gnome amd64 0.105-1ubuntu3.1 [27.7 kB]\nGet:153 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ttf-dejavu-extra all 2.33-2ubuntu1 [3,420 kB]\nGet:154 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe icedtea-7-jre-jamvm amd64 7u51-2.4.4-0ubuntu0.12.04.2 [558 kB]\nFetched 100 MB in 10s (9,721 kB/s)\nExtracting templates from packages: 19%\nExtracting templates from packages: 38%\nExtracting templates from packages: 58%\nExtracting templates from packages: 77%\nExtracting templates from packages: 97%\nExtracting templates from packages: 100%\nPreconfiguring packages ...\nSelecting previously unselected package libasound2.\n(Reading database ... \n(Reading database ... 5%\n(Reading database ... 10%\n(Reading database ... 15%\n(Reading database ... 20%\n(Reading database ... 25%\n(Reading database ... 30%\n(Reading database ... 35%\n(Reading database ... 40%\n(Reading database ... 45%\n(Reading database ... 50%\n(Reading database ... 55%\n(Reading database ... 60%\n(Reading database ... 65%\n(Reading database ... 70%\n(Reading database ... 75%\n(Reading database ... 80%\n(Reading database ... 85%\n(Reading database ... 90%\n(Reading database ... 95%\n(Reading database ... 100%\n(Reading database ... 47578 files and directories currently installed.)\nUnpacking libasound2 (from .../libasound2_1.0.25-1ubuntu10.2_amd64.deb) ...\nSelecting previously unselected package libasyncns0.\nUnpacking libasyncns0 (from .../libasyncns0_0.8-4_amd64.deb) ...\nSelecting previously unselected package libatasmart4.\nUnpacking libatasmart4 (from .../libatasmart4_0.18-3_amd64.deb) ...\nSelecting previously unselected package libatk1.0-data.\nUnpacking libatk1.0-data (from .../libatk1.0-data_2.4.0-0ubuntu1_all.deb) ...\nSelecting previously unselected package libatk1.0-0.\nUnpacking libatk1.0-0 (from .../libatk1.0-0_2.4.0-0ubuntu1_amd64.deb) ...\nSelecting previously unselected package libgtk2.0-common.\nUnpacking libgtk2.0-common (from .../libgtk2.0-common_2.24.10-0ubuntu6.1_all.deb) ...\nSelecting previously unselected package ttf-dejavu-core.\nUnpacking ttf-dejavu-core (from .../ttf-dejavu-core_2.33-2ubuntu1_all.deb) ...\nSelecting previously unselected package fontconfig-config.\nUnpacking fontconfig-config (from .../fontconfig-config_2.8.0-3ubuntu9.1_all.deb) ...\nSelecting previously unselected package libfontconfig1.\nUnpacking libfontconfig1 (from .../libfontconfig1_2.8.0-3ubuntu9.1_amd64.deb) ...\nSelecting previously unselected package libpixman-1-0.\nUnpacking libpixman-1-0 (from .../libpixman-1-0_0.30.2-1ubuntu0.0.0.0.1_amd64.deb) ...\nSelecting previously unselected package libxcb-render0.\nUnpacking libxcb-render0 (from .../libxcb-render0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxcb-shm0.\nUnpacking libxcb-shm0 (from .../libxcb-shm0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxrender1.\nUnpacking libxrender1 (from .../libxrender1_1%3a0.9.6-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libcairo2.\nUnpacking libcairo2 (from .../libcairo2_1.10.2-6.1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libavahi-common-data.\nUnpacking libavahi-common-data (from .../libavahi-common-data_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libavahi-common3.\nUnpacking libavahi-common3 (from .../libavahi-common3_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libavahi-client3.\nUnpacking libavahi-client3 (from .../libavahi-client3_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libcups2.\nUnpacking libcups2 (from .../libcups2_1.5.3-0ubuntu8.2_amd64.deb) ...\nSelecting previously unselected package libjpeg-turbo8.\nUnpacking libjpeg-turbo8 (from .../libjpeg-turbo8_1.1.90+svn733-0ubuntu4.3_amd64.deb) ...\nSelecting previously unselected package libjpeg8.\nUnpacking libjpeg8 (from .../libjpeg8_8c-2ubuntu7_amd64.deb) ...\nSelecting previously unselected package libjasper1.\nUnpacking libjasper1 (from .../libjasper1_1.900.1-13_amd64.deb) ...\nSelecting previously unselected package libtiff4.\nUnpacking libtiff4 (from .../libtiff4_3.9.5-2ubuntu1.5_amd64.deb) ...\nSelecting previously unselected package libgdk-pixbuf2.0-common.\nUnpacking libgdk-pixbuf2.0-common (from .../libgdk-pixbuf2.0-common_2.26.1-1ubuntu1.1_all.deb) ...\nSelecting previously unselected package libgdk-pixbuf2.0-0.\nUnpacking libgdk-pixbuf2.0-0 (from .../libgdk-pixbuf2.0-0_2.26.1-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libthai-data.\nUnpacking libthai-data (from .../libthai-data_0.1.16-3_all.deb) ...\nSelecting previously unselected package libdatrie1.\nUnpacking libdatrie1 (from .../libdatrie1_0.2.5-3_amd64.deb) ...\nSelecting previously unselected package libthai0.\nUnpacking libthai0 (from .../libthai0_0.1.16-3_amd64.deb) ...\nSelecting previously unselected package libxft2.\nUnpacking libxft2 (from .../libxft2_2.2.0-3ubuntu2_amd64.deb) ...\nSelecting previously unselected package fontconfig.\nUnpacking fontconfig (from .../fontconfig_2.8.0-3ubuntu9.1_amd64.deb) ...\nSelecting previously unselected package libpango1.0-0.\nUnpacking libpango1.0-0 (from .../libpango1.0-0_1.30.0-0ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package libxcomposite1.\nUnpacking libxcomposite1 (from .../libxcomposite1_1%3a0.4.3-2build1_amd64.deb) ...\nSelecting previously unselected package libxfixes3.\nUnpacking libxfixes3 (from .../libxfixes3_1%3a5.0-4ubuntu4.2_amd64.deb) ...\nSelecting previously unselected package libxcursor1.\nUnpacking libxcursor1 (from .../libxcursor1_1%3a1.1.12-1ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxdamage1.\nUnpacking libxdamage1 (from .../libxdamage1_1%3a1.1.3-2build1_amd64.deb) ...\nSelecting previously unselected package libxi6.\nUnpacking libxi6 (from .../libxi6_2%3a1.7.1.901-1ubuntu1~precise1_amd64.deb) ...\nSelecting previously unselected package libxinerama1.\nUnpacking libxinerama1 (from .../libxinerama1_2%3a1.1.1-3ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxrandr2.\nUnpacking libxrandr2 (from .../libxrandr2_2%3a1.3.2-2ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package shared-mime-info.\nUnpacking shared-mime-info (from .../shared-mime-info_1.0-0ubuntu4.1_amd64.deb) ...\nSelecting previously unselected package libgtk2.0-0.\nUnpacking libgtk2.0-0 (from .../libgtk2.0-0_2.24.10-0ubuntu6.1_amd64.deb) ...\nSelecting previously unselected package libnspr4.\nUnpacking libnspr4 (from .../libnspr4_4.9.5-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libnss3.\nUnpacking libnss3 (from .../libnss3_3.15.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libnss3-1d.\nUnpacking libnss3-1d (from .../libnss3-1d_3.15.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package ca-certificates-java.\nUnpacking ca-certificates-java (from .../ca-certificates-java_20110912ubuntu6_all.deb) ...\nSelecting previously unselected package tzdata-java.\nUnpacking tzdata-java (from .../tzdata-java_2014a-0ubuntu0.12.04_all.deb) ...\nSelecting previously unselected package java-common.\nUnpacking java-common (from .../java-common_0.43ubuntu2_all.deb) ...\nSelecting previously unselected package liblcms2-2.\nUnpacking liblcms2-2 (from .../liblcms2-2_2.2+git20110628-2ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jre-headless.\nUnpacking openjdk-7-jre-headless (from .../openjdk-7-jre-headless_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libgif4.\nUnpacking libgif4 (from .../libgif4_4.1.6-9ubuntu1_amd64.deb) ...\nSelecting previously unselected package libjson0.\nUnpacking libjson0 (from .../libjson0_0.9-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libogg0.\nUnpacking libogg0 (from .../libogg0_1.2.2~dfsg-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libflac8.\nUnpacking libflac8 (from .../libflac8_1.2.1-6_amd64.deb) ...\nSelecting previously unselected package libvorbis0a.\nUnpacking libvorbis0a (from .../libvorbis0a_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libvorbisenc2.\nUnpacking libvorbisenc2 (from .../libvorbisenc2_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libsndfile1.\nUnpacking libsndfile1 (from .../libsndfile1_1.0.25-4_amd64.deb) ...\nSelecting previously unselected package libpulse0.\nUnpacking libpulse0 (from .../libpulse0_1%3a1.1-0ubuntu15.4_amd64.deb) ...\nSelecting previously unselected package x11-common.\nUnpacking x11-common (from .../x11-common_1%3a7.6+12ubuntu2_all.deb) ...\nSelecting previously unselected package libxtst6.\nUnpacking libxtst6 (from .../libxtst6_2%3a1.2.0-4ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libglapi-mesa.\nUnpacking libglapi-mesa (from .../libglapi-mesa_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package libx11-xcb1.\nUnpacking libx11-xcb1 (from .../libx11-xcb1_2%3a1.4.99.1-0ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libxcb-glx0.\nUnpacking libxcb-glx0 (from .../libxcb-glx0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxxf86vm1.\nUnpacking libxxf86vm1 (from .../libxxf86vm1_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libgl1-mesa-glx.\nUnpacking libgl1-mesa-glx (from .../libgl1-mesa-glx_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jre.\nUnpacking openjdk-7-jre (from .../openjdk-7-jre_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libatk-wrapper-java.\nUnpacking libatk-wrapper-java (from .../libatk-wrapper-java_0.30.4-0ubuntu2_all.deb) ...\nSelecting previously unselected package libatk-wrapper-java-jni.\nUnpacking libatk-wrapper-java-jni (from .../libatk-wrapper-java-jni_0.30.4-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libavahi-glib1.\nUnpacking libavahi-glib1 (from .../libavahi-glib1_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libbonobo2-common.\nUnpacking libbonobo2-common (from .../libbonobo2-common_2.32.1-0ubuntu1.1_all.deb) ...\nSelecting previously unselected package libmpfr4.\nUnpacking libmpfr4 (from .../libmpfr4_3.1.0-3ubuntu2_amd64.deb) ...\nSelecting previously unselected package libmpc2.\nUnpacking libmpc2 (from .../libmpc2_0.9-4_amd64.deb) ...\nSelecting previously unselected package cpp-4.6.\nUnpacking cpp-4.6 (from .../cpp-4.6_4.6.3-1ubuntu5_amd64.deb) ...\nSelecting previously unselected package cpp.\nUnpacking cpp (from .../cpp_4%3a4.6.3-1ubuntu5_amd64.deb) ...\nSelecting previously unselected package libidl-common.\nUnpacking libidl-common (from .../libidl-common_0.8.14-0.2ubuntu2_all.deb) ...\nSelecting previously unselected package libidl0.\nUnpacking libidl0 (from .../libidl0_0.8.14-0.2ubuntu2_amd64.deb) ...\nSelecting previously unselected package liborbit2.\nUnpacking liborbit2 (from .../liborbit2_1%3a2.14.19-0.1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libbonobo2-0.\nUnpacking libbonobo2-0 (from .../libbonobo2-0_2.32.1-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libcairo-gobject2.\nUnpacking libcairo-gobject2 (from .../libcairo-gobject2_1.10.2-6.1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libltdl7.\nUnpacking libltdl7 (from .../libltdl7_2.4.2-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libtdb1.\nUnpacking libtdb1 (from .../libtdb1_1.2.9-4_amd64.deb) ...\nSelecting previously unselected package libvorbisfile3.\nUnpacking libvorbisfile3 (from .../libvorbisfile3_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package sound-theme-freedesktop.\nUnpacking sound-theme-freedesktop (from .../sound-theme-freedesktop_0.7.pristine-2_all.deb) ...\nSelecting previously unselected package libcanberra0.\nUnpacking libcanberra0 (from .../libcanberra0_0.28-3ubuntu3_amd64.deb) ...\nSelecting previously unselected package gconf2-common.\nUnpacking gconf2-common (from .../gconf2-common_3.2.5-0ubuntu2_all.deb) ...\nSelecting previously unselected package libgconf-2-4.\nUnpacking libgconf-2-4 (from .../libgconf-2-4_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgnome-keyring-common.\nUnpacking libgnome-keyring-common (from .../libgnome-keyring-common_3.2.2-2_all.deb) ...\nSelecting previously unselected package libgnome-keyring0.\nUnpacking libgnome-keyring0 (from .../libgnome-keyring0_3.2.2-2_amd64.deb) ...\nSelecting previously unselected package libgudev-1.0-0.\nUnpacking libgudev-1.0-0 (from .../libgudev-1.0-0_1%3a175-0ubuntu9.5_amd64.deb) ...\nSelecting previously unselected package libdevmapper-event1.02.1.\nUnpacking libdevmapper-event1.02.1 (from .../libdevmapper-event1.02.1_2%3a1.02.48-4ubuntu7.4_amd64.deb) ...\nSelecting previously unselected package liblvm2app2.2.\nUnpacking liblvm2app2.2 (from .../liblvm2app2.2_2.02.66-4ubuntu7.4_amd64.deb) ...\nSelecting previously unselected package libsgutils2-2.\nUnpacking libsgutils2-2 (from .../libsgutils2-2_1.33-1_amd64.deb) ...\nSelecting previously unselected package udisks.\nUnpacking udisks (from .../udisks_1.0.4-5ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libgdu0.\nUnpacking libgdu0 (from .../libgdu0_3.0.2-2ubuntu7_amd64.deb) ...\nSelecting previously unselected package gconf-service-backend.\nUnpacking gconf-service-backend (from .../gconf-service-backend_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package gconf-service.\nUnpacking gconf-service (from .../gconf-service_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package dbus-x11.\nUnpacking dbus-x11 (from .../dbus-x11_1.4.18-1ubuntu1.4_amd64.deb) ...\nSelecting previously unselected package gconf2.\nUnpacking gconf2 (from .../gconf2_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgnomevfs2-common.\nUnpacking libgnomevfs2-common (from .../libgnomevfs2-common_1%3a2.24.4-1ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libgnomevfs2-0.\nUnpacking libgnomevfs2-0 (from .../libgnomevfs2-0_1%3a2.24.4-1ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libgnome2-common.\nUnpacking libgnome2-common (from .../libgnome2-common_2.32.1-2ubuntu1.1_all.deb) ...\nSelecting previously unselected package libgnome2-bin.\nUnpacking libgnome2-bin (from .../libgnome2-bin_2.32.1-2ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libgnome2-0.\nUnpacking libgnome2-0 (from .../libgnome2-0_2.32.1-2ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libdconf0.\nUnpacking libdconf0 (from .../libdconf0_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package dconf-service.\nUnpacking dconf-service (from .../dconf-service_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package dconf-gsettings-backend.\nUnpacking dconf-gsettings-backend (from .../dconf-gsettings-backend_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libgtk-3-common.\nUnpacking libgtk-3-common (from .../libgtk-3-common_3.4.2-0ubuntu0.8_all.deb) ...\nSelecting previously unselected package libgtk-3-0.\nUnpacking libgtk-3-0 (from .../libgtk-3-0_3.4.2-0ubuntu0.8_amd64.deb) ...\nSelecting previously unselected package libice6.\nUnpacking libice6 (from .../libice6_2%3a1.0.7-2build1_amd64.deb) ...\nSelecting previously unselected package libllvm3.0.\nUnpacking libllvm3.0 (from .../libllvm3.0_3.0-4ubuntu1_amd64.deb) ...\nSelecting previously unselected package libpolkit-agent-1-0.\nUnpacking libpolkit-agent-1-0 (from .../libpolkit-agent-1-0_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libpolkit-backend-1-0.\nUnpacking libpolkit-backend-1-0 (from .../libpolkit-backend-1-0_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libsm6.\nUnpacking libsm6 (from .../libsm6_2%3a1.2.0-2build1_amd64.deb) ...\nSelecting previously unselected package libxt6.\nUnpacking libxt6 (from .../libxt6_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxmu6.\nUnpacking libxmu6 (from .../libxmu6_2%3a1.1.0-3_amd64.deb) ...\nSelecting previously unselected package libxpm4.\nUnpacking libxpm4 (from .../libxpm4_1%3a3.5.9-4_amd64.deb) ...\nSelecting previously unselected package libxaw7.\nUnpacking libxaw7 (from .../libxaw7_2%3a1.0.9-3ubuntu1_amd64.deb) ...\nSelecting previously unselected package libxcb-shape0.\nUnpacking libxcb-shape0 (from .../libxcb-shape0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxv1.\nUnpacking libxv1 (from .../libxv1_2%3a1.0.6-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libck-connector0.\nUnpacking libck-connector0 (from .../libck-connector0_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package consolekit.\nUnpacking consolekit (from .../consolekit_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libcryptsetup4.\nUnpacking libcryptsetup4 (from .../libcryptsetup4_2%3a1.4.1-2ubuntu4_amd64.deb) ...\nSelecting previously unselected package cryptsetup-bin.\nUnpacking cryptsetup-bin (from .../cryptsetup-bin_2%3a1.4.1-2ubuntu4_amd64.deb) ...\nSelecting previously unselected package libfontenc1.\nUnpacking libfontenc1 (from .../libfontenc1_1%3a1.1.0-1_amd64.deb) ...\nSelecting previously unselected package libxxf86dga1.\nUnpacking libxxf86dga1 (from .../libxxf86dga1_2%3a1.1.2-1ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package x11-utils.\nUnpacking x11-utils (from .../x11-utils_7.6+4ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package gvfs-common.\nUnpacking gvfs-common (from .../gvfs-common_1.12.1-0ubuntu1.2_all.deb) ...\nSelecting previously unselected package gvfs-libs.\nUnpacking gvfs-libs (from .../gvfs-libs_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package gvfs-daemons.\nUnpacking gvfs-daemons (from .../gvfs-daemons_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package gvfs.\nUnpacking gvfs (from .../gvfs_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package hicolor-icon-theme.\nUnpacking hicolor-icon-theme (from .../hicolor-icon-theme_0.12-1ubuntu2_all.deb) ...\nSelecting previously unselected package libgconf2-4.\nUnpacking libgconf2-4 (from .../libgconf2-4_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgl1-mesa-dri.\nUnpacking libgl1-mesa-dri (from .../libgl1-mesa-dri_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package libgtk-3-bin.\nUnpacking libgtk-3-bin (from .../libgtk-3-bin_3.4.2-0ubuntu0.8_amd64.deb) ...\nAdding 'diversion of /usr/sbin/update-icon-caches to /usr/sbin/update-icon-caches.gtk2 by libgtk-3-bin'\nAdding 'diversion of /usr/share/man/man8/update-icon-caches.8.gz to /usr/share/man/man8/update-icon-caches.gtk2.8.gz by libgtk-3-bin'\nSelecting previously unselected package libgtk2.0-bin.\nUnpacking libgtk2.0-bin (from .../libgtk2.0-bin_2.24.10-0ubuntu6.1_amd64.deb) ...\nSelecting previously unselected package xorg-sgml-doctools.\nUnpacking xorg-sgml-doctools (from .../xorg-sgml-doctools_1%3a1.10-1_all.deb) ...\nSelecting previously unselected package x11proto-core-dev.\nUnpacking x11proto-core-dev (from .../x11proto-core-dev_7.0.22-1ubuntu0.1_all.deb) ...\nSelecting previously unselected package libice-dev.\nUnpacking libice-dev (from .../libice-dev_2%3a1.0.7-2build1_amd64.deb) ...\nSelecting previously unselected package libpam-ck-connector.\nUnpacking libpam-ck-connector (from .../libpam-ck-connector_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libpthread-stubs0.\nUnpacking libpthread-stubs0 (from .../libpthread-stubs0_0.3-3_amd64.deb) ...\nSelecting previously unselected package libpthread-stubs0-dev.\nUnpacking libpthread-stubs0-dev (from .../libpthread-stubs0-dev_0.3-3_amd64.deb) ...\nSelecting previously unselected package libsm-dev.\nUnpacking libsm-dev (from .../libsm-dev_2%3a1.2.0-2build1_amd64.deb) ...\nSelecting previously unselected package libxau-dev.\nUnpacking libxau-dev (from .../libxau-dev_1%3a1.0.6-4_amd64.deb) ...\nSelecting previously unselected package libxdmcp-dev.\nUnpacking libxdmcp-dev (from .../libxdmcp-dev_1%3a1.1.0-4_amd64.deb) ...\nSelecting previously unselected package x11proto-input-dev.\nUnpacking x11proto-input-dev (from .../x11proto-input-dev_2.3-1~precise1_all.deb) ...\nSelecting previously unselected package x11proto-kb-dev.\nUnpacking x11proto-kb-dev (from .../x11proto-kb-dev_1.0.5-2_all.deb) ...\nSelecting previously unselected package xtrans-dev.\nUnpacking xtrans-dev (from .../xtrans-dev_1.2.6-2_all.deb) ...\nSelecting previously unselected package libxcb1-dev.\nUnpacking libxcb1-dev (from .../libxcb1-dev_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libx11-dev.\nUnpacking libx11-dev (from .../libx11-dev_2%3a1.4.99.1-0ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libx11-doc.\nUnpacking libx11-doc (from .../libx11-doc_2%3a1.4.99.1-0ubuntu2.2_all.deb) ...\nSelecting previously unselected package libxt-dev.\nUnpacking libxt-dev (from .../libxt-dev_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package mtools.\nUnpacking mtools (from .../mtools_4.0.12-1ubuntu0.12.04.1_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jdk.\nUnpacking openjdk-7-jdk (from .../openjdk-7-jdk_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package policykit-1.\nUnpacking policykit-1 (from .../policykit-1_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package policykit-1-gnome.\nUnpacking policykit-1-gnome (from .../policykit-1-gnome_0.105-1ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package ttf-dejavu-extra.\nUnpacking ttf-dejavu-extra (from .../ttf-dejavu-extra_2.33-2ubuntu1_all.deb) ...\nSelecting previously unselected package icedtea-7-jre-jamvm.\nUnpacking icedtea-7-jre-jamvm (from .../icedtea-7-jre-jamvm_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nProcessing triggers for initramfs-tools ...\nupdate-initramfs: Generating /boot/initrd.img-3.2.0-60-virtual\nProcessing triggers for man-db ...\nProcessing triggers for ureadahead ...\nProcessing triggers for libglib2.0-0 ...\nProcessing triggers for install-info ...\nSetting up libasound2 (1.0.25-1ubuntu10.2) ...\nSetting up libasyncns0 (0.8-4) ...\nSetting up libatasmart4 (0.18-3) ...\nSetting up libatk1.0-data (2.4.0-0ubuntu1) ...\nSetting up libatk1.0-0 (2.4.0-0ubuntu1) ...\nSetting up libgtk2.0-common (2.24.10-0ubuntu6.1) ...\nSetting up ttf-dejavu-core (2.33-2ubuntu1) ...\nSetting up fontconfig-config (2.8.0-3ubuntu9.1) ...\nSetting up libfontconfig1 (2.8.0-3ubuntu9.1) ...\nSetting up libpixman-1-0 (0.30.2-1ubuntu0.0.0.0.1) ...\nSetting up libxcb-render0 (1.8.1-1ubuntu0.2) ...\nSetting up libxcb-shm0 (1.8.1-1ubuntu0.2) ...\nSetting up libxrender1 (1:0.9.6-2ubuntu0.1) ...\nSetting up libcairo2 (1.10.2-6.1ubuntu3) ...\nSetting up libavahi-common-data (0.6.30-5ubuntu2.1) ...\nSetting up libavahi-common3 (0.6.30-5ubuntu2.1) ...\nSetting up libavahi-client3 (0.6.30-5ubuntu2.1) ...\nSetting up libcups2 (1.5.3-0ubuntu8.2) ...\nSetting up libjpeg-turbo8 (1.1.90+svn733-0ubuntu4.3) ...\nSetting up libjpeg8 (8c-2ubuntu7) ...\nSetting up libjasper1 (1.900.1-13) ...\nSetting up libtiff4 (3.9.5-2ubuntu1.5) ...\nSetting up libgdk-pixbuf2.0-common (2.26.1-1ubuntu1.1) ...\nSetting up libgdk-pixbuf2.0-0 (2.26.1-1ubuntu1.1) ...\nSetting up libthai-data (0.1.16-3) ...\nSetting up libdatrie1 (0.2.5-3) ...\nSetting up libthai0 (0.1.16-3) ...\nSetting up libxft2 (2.2.0-3ubuntu2) ...\nSetting up fontconfig (2.8.0-3ubuntu9.1) ...\nCleaning up old fontconfig caches... done.\nRegenerating fonts cache... done.\nSetting up libpango1.0-0 (1.30.0-0ubuntu3.1) ...\nSetting up libxcomposite1 (1:0.4.3-2build1) ...\nSetting up libxfixes3 (1:5.0-4ubuntu4.2) ...\nSetting up libxcursor1 (1:1.1.12-1ubuntu0.1) ...\nSetting up libxdamage1 (1:1.1.3-2build1) ...\nSetting up libxi6 (2:1.7.1.901-1ubuntu1~precise1) ...\nSetting up libxinerama1 (2:1.1.1-3ubuntu0.1) ...\nSetting up libxrandr2 (2:1.3.2-2ubuntu0.2) ...\nSetting up shared-mime-info (1.0-0ubuntu4.1) ...\nSetting up libgtk2.0-0 (2.24.10-0ubuntu6.1) ...\nSetting up libnspr4 (4.9.5-0ubuntu0.12.04.2) ...\nSetting up libnss3 (3.15.4-0ubuntu0.12.04.2) ...\nSetting up libnss3-1d (3.15.4-0ubuntu0.12.04.2) ...\nSetting up tzdata-java (2014a-0ubuntu0.12.04) ...\nSetting up java-common (0.43ubuntu2) ...\nSetting up liblcms2-2 (2.2+git20110628-2ubuntu3.1) ...\nSetting up libgif4 (4.1.6-9ubuntu1) ...\nSetting up libjson0 (0.9-1ubuntu1) ...\nSetting up libogg0 (1.2.2~dfsg-1ubuntu1) ...\nSetting up libflac8 (1.2.1-6) ...\nSetting up libvorbis0a (1.3.2-1ubuntu3) ...\nSetting up libvorbisenc2 (1.3.2-1ubuntu3) ...\nSetting up libsndfile1 (1.0.25-4) ...\nSetting up libpulse0 (1:1.1-0ubuntu15.4) ...\nSetting up x11-common (1:7.6+12ubuntu2) ...\nSetting up libxtst6 (2:1.2.0-4ubuntu0.1) ...\nSetting up libglapi-mesa (8.0.4-0ubuntu0.7) ...\nSetting up libx11-xcb1 (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libxcb-glx0 (1.8.1-1ubuntu0.2) ...\nSetting up libxxf86vm1 (1:1.1.1-2ubuntu0.1) ...\nSetting up libgl1-mesa-glx (8.0.4-0ubuntu0.7) ...\nupdate-alternatives: using /usr/lib/x86_64-linux-gnu/mesa/ld.so.conf to provide /etc/ld.so.conf.d/x86_64-linux-gnu_GL.conf (x86_64-linux-gnu_gl_conf) in auto mode.\nSetting up libavahi-glib1 (0.6.30-5ubuntu2.1) ...\nSetting up libbonobo2-common (2.32.1-0ubuntu1.1) ...\nSetting up libmpfr4 (3.1.0-3ubuntu2) ...\nSetting up libmpc2 (0.9-4) ...\nSetting up cpp-4.6 (4.6.3-1ubuntu5) ...\nSetting up cpp (4:4.6.3-1ubuntu5) ...\nSetting up libidl-common (0.8.14-0.2ubuntu2) ...\nSetting up libidl0 (0.8.14-0.2ubuntu2) ...\nSetting up liborbit2 (1:2.14.19-0.1ubuntu1) ...\nSetting up libbonobo2-0 (2.32.1-0ubuntu1.1) ...\nSetting up libcairo-gobject2 (1.10.2-6.1ubuntu3) ...\nSetting up libltdl7 (2.4.2-1ubuntu1) ...\nSetting up libtdb1 (1.2.9-4) ...\nSetting up libvorbisfile3 (1.3.2-1ubuntu3) ...\nSetting up sound-theme-freedesktop (0.7.pristine-2) ...\nSetting up libcanberra0 (0.28-3ubuntu3) ...\nSetting up gconf2-common (3.2.5-0ubuntu2) ...\nCreating config file /etc/gconf/2/path with new version\nSetting up libgconf-2-4 (3.2.5-0ubuntu2) ...\nSetting up libgnome-keyring-common (3.2.2-2) ...\nSetting up libgnome-keyring0 (3.2.2-2) ...\nSetting up libgudev-1.0-0 (1:175-0ubuntu9.5) ...\nSetting up libdevmapper-event1.02.1 (2:1.02.48-4ubuntu7.4) ...\nSetting up liblvm2app2.2 (2.02.66-4ubuntu7.4) ...\nSetting up libsgutils2-2 (1.33-1) ...\nSetting up udisks (1.0.4-5ubuntu2.2) ...\nSetting up libgdu0 (3.0.2-2ubuntu7) ...\nSetting up dbus-x11 (1.4.18-1ubuntu1.4) ...\nSetting up libdconf0 (0.12.0-0ubuntu1.1) ...\nSetting up dconf-service (0.12.0-0ubuntu1.1) ...\nSetting up dconf-gsettings-backend (0.12.0-0ubuntu1.1) ...\nSetting up libgtk-3-common (3.4.2-0ubuntu0.8) ...\nSetting up libgtk-3-0 (3.4.2-0ubuntu0.8) ...\nSetting up libice6 (2:1.0.7-2build1) ...\nSetting up libllvm3.0 (3.0-4ubuntu1) ...\nSetting up libpolkit-agent-1-0 (0.104-1ubuntu1.1) ...\nSetting up libpolkit-backend-1-0 (0.104-1ubuntu1.1) ...\nSetting up libsm6 (2:1.2.0-2build1) ...\nSetting up libxt6 (1:1.1.1-2ubuntu0.1) ...\nSetting up libxmu6 (2:1.1.0-3) ...\nSetting up libxpm4 (1:3.5.9-4) ...\nSetting up libxaw7 (2:1.0.9-3ubuntu1) ...\nSetting up libxcb-shape0 (1.8.1-1ubuntu0.2) ...\nSetting up libxv1 (2:1.0.6-2ubuntu0.1) ...\nSetting up libck-connector0 (0.4.5-2ubuntu0.1) ...\nSetting up consolekit (0.4.5-2ubuntu0.1) ...\nSetting up libcryptsetup4 (2:1.4.1-2ubuntu4) ...\nSetting up cryptsetup-bin (2:1.4.1-2ubuntu4) ...\nSetting up libfontenc1 (1:1.1.0-1) ...\nSetting up libxxf86dga1 (2:1.1.2-1ubuntu0.1) ...\nSetting up x11-utils (7.6+4ubuntu0.1) ...\nSetting up gvfs-common (1.12.1-0ubuntu1.2) ...\nSetting up gvfs-libs (1.12.1-0ubuntu1.2) ...\nSetting up gvfs-daemons (1.12.1-0ubuntu1.2) ...\nSetting up gvfs (1.12.1-0ubuntu1.2) ...\nSetting up hicolor-icon-theme (0.12-1ubuntu2) ...\nSetting up libgl1-mesa-dri (8.0.4-0ubuntu0.7) ...\nSetting up libgtk-3-bin (3.4.2-0ubuntu0.8) ...\nSetting up libgtk2.0-bin (2.24.10-0ubuntu6.1) ...\nSetting up xorg-sgml-doctools (1:1.10-1) ...\nSetting up x11proto-core-dev (7.0.22-1ubuntu0.1) ...\nSetting up libice-dev (2:1.0.7-2build1) ...\nSetting up libpam-ck-connector (0.4.5-2ubuntu0.1) ...\nSetting up libpthread-stubs0 (0.3-3) ...\nSetting up libpthread-stubs0-dev (0.3-3) ...\nSetting up libsm-dev (2:1.2.0-2build1) ...\nSetting up libxau-dev (1:1.0.6-4) ...\nSetting up libxdmcp-dev (1:1.1.0-4) ...\nSetting up x11proto-input-dev (2.3-1~precise1) ...\nSetting up x11proto-kb-dev (1.0.5-2) ...\nSetting up xtrans-dev (1.2.6-2) ...\nSetting up libxcb1-dev (1.8.1-1ubuntu0.2) ...\nSetting up libx11-dev (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libx11-doc (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libxt-dev (1:1.1.1-2ubuntu0.1) ...\nSetting up mtools (4.0.12-1ubuntu0.12.04.1) ...\nSetting up policykit-1 (0.104-1ubuntu1.1) ...\nSetting up policykit-1-gnome (0.105-1ubuntu3.1) ...\nSetting up ttf-dejavu-extra (2.33-2ubuntu1) ...\nSetting up gconf-service (3.2.5-0ubuntu2) ...\nSetting up libgconf2-4 (3.2.5-0ubuntu2) ...\nSetting up openjdk-7-jre-headless (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/java to provide /usr/bin/java (java) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/keytool to provide /usr/bin/keytool (keytool) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/pack200 to provide /usr/bin/pack200 (pack200) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/rmid to provide /usr/bin/rmid (rmid) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/rmiregistry to provide /usr/bin/rmiregistry (rmiregistry) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/unpack200 to provide /usr/bin/unpack200 (unpack200) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/orbd to provide /usr/bin/orbd (orbd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/servertool to provide /usr/bin/servertool (servertool) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/tnameserv to provide /usr/bin/tnameserv (tnameserv) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/lib/jexec to provide /usr/bin/jexec (jexec) in auto mode.\nSetting up libatk-wrapper-java (0.30.4-0ubuntu2) ...\nSetting up gconf-service-backend (3.2.5-0ubuntu2) ...\nSetting up gconf2 (3.2.5-0ubuntu2) ...\nupdate-alternatives: using /usr/bin/gconftool-2 to provide /usr/bin/gconftool (gconftool) in auto mode.\nSetting up libgnomevfs2-common (1:2.24.4-1ubuntu2.1) ...\nSetting up libgnomevfs2-0 (1:2.24.4-1ubuntu2.1) ...\nSetting up libgnome2-common (2.32.1-2ubuntu1.1) ...\nSetting up libgnome2-bin (2.32.1-2ubuntu1.1) ...\nSetting up libgnome2-0 (2.32.1-2ubuntu1.1) ...\nSetting up icedtea-7-jre-jamvm (7u51-2.4.4-0ubuntu0.12.04.2) ...\nSetting up ca-certificates-java (20110912ubuntu6) ...\nAdding debian:Digital_Signature_Trust_Co._Global_CA_1.pem\nAdding debian:Wells_Fargo_Root_CA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority.pem\nAdding debian:Staat_der_Nederlanden_Root_CA.pem\nAdding debian:ACEDICOM_Root.pem\nAdding debian:Secure_Global_CA.pem\nAdding debian:Entrust.net_Premium_2048_Secure_Server_CA.pem\nAdding debian:NetLock_Business_=Class_B=_Root.pem\nAdding debian:Comodo_Secure_Services_root.pem\nAdding debian:AffirmTrust_Premium.pem\nAdding debian:StartCom_Certification_Authority_G2.pem\nAdding debian:GlobalSign_Root_CA_-_R2.pem\nAdding debian:TC_TrustCenter_Class_2_CA_II.pem\nAdding debian:China_Internet_Network_Information_Center_EV_Certificates_Root.pem\nAdding debian:SwissSign_Gold_CA_-_G2.pem\nAdding debian:NetLock_Notary_=Class_A=_Root.pem\nAdding debian:AddTrust_Public_Services_Root.pem\nAdding debian:GeoTrust_Primary_Certification_Authority.pem\nAdding debian:Starfield_Root_Certificate_Authority_-_G2.pem\nAdding debian:DigiCert_Global_Root_CA.pem\nAdding debian:Firmaprofesional_Root_CA.pem\nAdding debian:D-TRUST_Root_Class_3_CA_2_2009.pem\nAdding debian:GeoTrust_Universal_CA_2.pem\nAdding debian:DigiCert_High_Assurance_EV_Root_CA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:VeriSign_Class_3_Public_Primary_Certification_Authority_-_G4.pem\nAdding debian:Izenpe.com.pem\nAdding debian:Verisign_Class_2_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:A-Trust-nQual-03.pem\nAdding debian:Swisscom_Root_CA_1.pem\nAdding debian:Buypass_Class_2_CA_1.pem\nAdding debian:CA_Disig_Root_R2.pem\nAdding debian:QuoVadis_Root_CA.pem\nAdding debian:Network_Solutions_Certificate_Authority.pem\nAdding debian:GlobalSign_Root_CA_-_R3.pem\nAdding debian:Verisign_Class_4_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:Security_Communication_RootCA2.pem\nAdding debian:AddTrust_External_Root.pem\nAdding debian:Hellenic_Academic_and_Research_Institutions_RootCA_2011.pem\nAdding debian:COMODO_Certification_Authority.pem\nAdding debian:QuoVadis_Root_CA_2.pem\nAdding debian:GeoTrust_Global_CA.pem\nAdding debian:thawte_Primary_Root_CA_-_G3.pem\nAdding debian:Swisscom_Root_CA_2.pem\nAdding debian:Microsec_e-Szigno_Root_CA.pem\nAdding debian:UTN_USERFirst_Email_Root_CA.pem\nAdding debian:SwissSign_Platinum_CA_-_G2.pem\nAdding debian:Certum_Root_CA.pem\nAdding debian:spi-cacert-2008.pem\nAdding debian:Thawte_Premium_Server_CA.pem\nAdding debian:Global_Chambersign_Root_-_2008.pem\nAdding debian:Root_CA_Generalitat_Valenciana.pem\nAdding debian:GTE_CyberTrust_Global_Root.pem\nAdding debian:SecureTrust_CA.pem\nAdding debian:GeoTrust_Global_CA_2.pem\nAdding debian:NetLock_Qualified_=Class_QA=_Root.pem\nAdding debian:EE_Certification_Centre_Root_CA.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:COMODO_ECC_Certification_Authority.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_2007.pem\nAdding debian:ePKI_Root_Certification_Authority.pem\nAdding debian:E-Guven_Kok_Elektronik_Sertifika_Hizmet_Saglayicisi.pem\nAdding debian:WellsSecure_Public_Root_Certificate_Authority.pem\nAdding debian:TDC_Internet_Root_CA.pem\nAdding debian:Juur-SK.pem\nAdding debian:DigiCert_Assured_ID_Root_CA.pem\nAdding debian:certSIGN_ROOT_CA.pem\nAdding debian:Starfield_Services_Root_Certificate_Authority_-_G2.pem\nAdding debian:EBG_Elektronik_Sertifika_Hizmet_Sağlayıcısı.pem\nAdding debian:AC_Raíz_Certicámara_S.A..pem\nAdding debian:XRamp_Global_CA_Root.pem\nAdding debian:America_Online_Root_Certification_Authority_2.pem\nAdding debian:GeoTrust_Universal_CA.pem\nAdding debian:GlobalSign_Root_CA.pem\nAdding debian:Visa_eCommerce_Root.pem\nAdding debian:Actalis_Authentication_Root_CA.pem\nAdding debian:America_Online_Root_Certification_Authority_1.pem\nAdding debian:NetLock_Express_=Class_C=_Root.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_2.pem\nAdding debian:D-TRUST_Root_Class_3_CA_2_EV_2009.pem\nAdding debian:ApplicationCA_-_Japanese_Government.pem\nAdding debian:QuoVadis_Root_CA_3.pem\nAdding debian:SwissSign_Silver_CA_-_G2.pem\nAdding debian:SecureSign_RootCA11.pem\nAdding debian:Deutsche_Telekom_Root_CA_2.pem\nAdding debian:GeoTrust_Primary_Certification_Authority_-_G3.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:thawte_Primary_Root_CA.pem\nAdding debian:Certigna.pem\nAdding debian:DST_Root_CA_X3.pem\nAdding debian:Baltimore_CyberTrust_Root.pem\nAdding debian:RSA_Root_Certificate_1.pem\nAdding debian:IGC_A.pem\nAdding debian:Trustis_FPS_Root_CA.pem\nAdding debian:PSCProcert.pem\nAdding debian:CA_Disig.pem\nAdding debian:Taiwan_GRCA.pem\nAdding debian:Chambers_of_Commerce_Root_-_2008.pem\nAdding debian:Verisign_Class_2_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:cert-ec2.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_1.pem\nAdding debian:Sonera_Class_2_Root_CA.pem\nAdding debian:DST_ACES_CA_X6.pem\nAdding debian:EC-ACC.pem\nAdding debian:OISTE_WISeKey_Global_Root_GA_CA.pem\nAdding debian:AddTrust_Qualified_Certificates_Root.pem\nAdding debian:Entrust_Root_Certification_Authority.pem\nAdding debian:thawte_Primary_Root_CA_-_G2.pem\nAdding debian:AffirmTrust_Commercial.pem\nAdding debian:Cybertrust_Global_Root.pem\nAdding debian:TDC_OCES_Root_CA.pem\nAdding debian:Sonera_Class_1_Root_CA.pem\nAdding debian:ComSign_Secured_CA.pem\nAdding debian:Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem\nAdding debian:S-TRUST_Authentication_and_Encryption_Root_CA_2005_PN.pem\nAdding debian:Go_Daddy_Class_2_CA.pem\nAdding debian:Comodo_Trusted_Services_root.pem\nAdding debian:Buypass_Class_3_CA_1.pem\nAdding debian:AffirmTrust_Networking.pem\nAdding debian:StartCom_Certification_Authority_2.pem\nAdding debian:Equifax_Secure_Global_eBusiness_CA.pem\nAdding debian:Digital_Signature_Trust_Co._Global_CA_3.pem\nAdding debian:Equifax_Secure_eBusiness_CA_1.pem\nAdding debian:Go_Daddy_Root_Certificate_Authority_-_G2.pem\nAdding debian:AffirmTrust_Premium_ECC.pem\nAdding debian:Microsec_e-Szigno_Root_CA_2009.pem\nAdding debian:Camerfirma_Global_Chambersign_Root.pem\nAdding debian:Comodo_AAA_Services_root.pem\nAdding debian:Swisscom_Root_EV_CA_2.pem\nAdding debian:VeriSign_Universal_Root_Certification_Authority.pem\nAdding debian:VeriSign_Class_3_Public_Primary_Certification_Authority_-_G5.pem\nAdding debian:Equifax_Secure_CA.pem\nAdding debian:Security_Communication_Root_CA.pem\nAdding debian:Entrust.net_Secure_Server_CA.pem\nAdding debian:Buypass_Class_2_Root_CA.pem\nAdding debian:ComSign_CA.pem\nAdding debian:GeoTrust_Primary_Certification_Authority_-_G2.pem\nAdding debian:TÜBİTAK_UEKAE_Kök_Sertifika_Hizmet_Sağlayıcısı_-_Sürüm_3.pem\nAdding debian:Certum_Trusted_Network_CA.pem\nAdding debian:TC_TrustCenter_Class_3_CA_II.pem\nAdding debian:ValiCert_Class_1_VA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:NetLock_Arany_=Class_Gold=_Főtanúsítvány.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_2.pem\nAdding debian:Certplus_Class_2_Primary_CA.pem\nAdding debian:Camerfirma_Chambers_of_Commerce_Root.pem\nAdding debian:Security_Communication_EV_RootCA1.pem\nAdding debian:StartCom_Certification_Authority.pem\nAdding debian:Thawte_Server_CA.pem\nAdding debian:RSA_Security_2048_v3.pem\nAdding debian:AddTrust_Low-Value_Services_Root.pem\nAdding debian:Certinomis_-_Autorité_Racine.pem\nAdding debian:Hongkong_Post_Root_CA_1.pem\nAdding debian:ValiCert_Class_2_VA.pem\nAdding debian:Staat_der_Nederlanden_Root_CA_-_G2.pem\nAdding debian:T-TeleSec_GlobalRoot_Class_3.pem\nAdding debian:UTN_DATACorp_SGC_Root_CA.pem\nAdding debian:TC_TrustCenter_Universal_CA_I.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority.pem\nAdding debian:TWCA_Root_Certification_Authority.pem\nAdding debian:Buypass_Class_3_Root_CA.pem\nAdding debian:CA_Disig_Root_R1.pem\nAdding debian:Starfield_Class_2_CA.pem\nAdding debian:CNNIC_ROOT.pem\nAdding debian:UTN_USERFirst_Hardware_Root_CA.pem\ndone.\nSetting up libatk-wrapper-java-jni (0.30.4-0ubuntu2) ...\nSetting up openjdk-7-jre (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/policytool to provide /usr/bin/policytool (policytool) in auto mode.\nSetting up openjdk-7-jdk (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/appletviewer to provide /usr/bin/appletviewer (appletviewer) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/extcheck to provide /usr/bin/extcheck (extcheck) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/idlj to provide /usr/bin/idlj (idlj) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jar to provide /usr/bin/jar (jar) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jarsigner to provide /usr/bin/jarsigner (jarsigner) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javac to provide /usr/bin/javac (javac) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javadoc to provide /usr/bin/javadoc (javadoc) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javah to provide /usr/bin/javah (javah) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javap to provide /usr/bin/javap (javap) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jcmd to provide /usr/bin/jcmd (jcmd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jconsole to provide /usr/bin/jconsole (jconsole) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jdb to provide /usr/bin/jdb (jdb) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jhat to provide /usr/bin/jhat (jhat) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jinfo to provide /usr/bin/jinfo (jinfo) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jmap to provide /usr/bin/jmap (jmap) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jps to provide /usr/bin/jps (jps) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jrunscript to provide /usr/bin/jrunscript (jrunscript) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jsadebugd to provide /usr/bin/jsadebugd (jsadebugd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstack to provide /usr/bin/jstack (jstack) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstat to provide /usr/bin/jstat (jstat) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstatd to provide /usr/bin/jstatd (jstatd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/native2ascii to provide /usr/bin/native2ascii (native2ascii) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/rmic to provide /usr/bin/rmic (rmic) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/schemagen to provide /usr/bin/schemagen (schemagen) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/serialver to provide /usr/bin/serialver (serialver) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/wsgen to provide /usr/bin/wsgen (wsgen) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/wsimport to provide /usr/bin/wsimport (wsimport) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/xjc to provide /usr/bin/xjc (xjc) in auto mode.\nProcessing triggers for libc-bin ...\nldconfig deferred processing now taking place\naccountsservice\t\t\t\t\tinstall\nacpid\t\t\t\t\t\tinstall\nadduser\t\t\t\t\t\tinstall\napparmor\t\t\t\t\tinstall\napport\t\t\t\t\t\tinstall\napport-symptoms\t\t\t\t\tinstall\napt\t\t\t\t\t\tinstall\napt-transport-https\t\t\t\tinstall\napt-utils\t\t\t\t\tinstall\napt-xapian-index\t\t\t\tinstall\naptitude\t\t\t\t\tinstall\nat\t\t\t\t\t\tinstall\nbase-files\t\t\t\t\tinstall\nbase-passwd\t\t\t\t\tinstall\nbash\t\t\t\t\t\tinstall\nbash-completion\t\t\t\t\tinstall\nbc\t\t\t\t\t\tinstall\nbind9-host\t\t\t\t\tinstall\nbsdmainutils\t\t\t\t\tinstall\nbsdutils\t\t\t\t\tinstall\nbusybox-initramfs\t\t\t\tinstall\nbusybox-static\t\t\t\t\tinstall\nbyobu\t\t\t\t\t\tinstall\nbzip2\t\t\t\t\t\tinstall\nca-certificates\t\t\t\t\tinstall\nca-certificates-java\t\t\t\tinstall\ncloud-init\t\t\t\t\tinstall\ncloud-initramfs-growroot\t\t\tinstall\ncloud-initramfs-rescuevol\t\t\tinstall\ncloud-utils\t\t\t\t\tinstall\ncommand-not-found\t\t\t\tinstall\ncommand-not-found-data\t\t\t\tinstall\nconsole-setup\t\t\t\t\tinstall\nconsolekit\t\t\t\t\tinstall\ncoreutils\t\t\t\t\tinstall\ncpio\t\t\t\t\t\tinstall\ncpp\t\t\t\t\t\tinstall\ncpp-4.6\t\t\t\t\t\tinstall\ncrda\t\t\t\t\t\tinstall\ncron\t\t\t\t\t\tinstall\ncryptsetup-bin\t\t\t\t\tinstall\ncurl\t\t\t\t\t\tinstall\ndash\t\t\t\t\t\tinstall\ndbus\t\t\t\t\t\tinstall\ndbus-x11\t\t\t\t\tinstall\ndconf-gsettings-backend\t\t\t\tinstall\ndconf-service\t\t\t\t\tinstall\ndebconf\t\t\t\t\t\tinstall\ndebconf-i18n\t\t\t\t\tinstall\ndebianutils\t\t\t\t\tinstall\ndiffutils\t\t\t\t\tinstall\ndmidecode\t\t\t\t\tinstall\ndmsetup\t\t\t\t\t\tinstall\ndnsutils\t\t\t\t\tinstall\ndosfstools\t\t\t\t\tinstall\ndpkg\t\t\t\t\t\tinstall\ne2fslibs\t\t\t\t\tinstall\ne2fsprogs\t\t\t\t\tinstall\ned\t\t\t\t\t\tinstall\neject\t\t\t\t\t\tinstall\neuca2ools\t\t\t\t\tinstall\nfile\t\t\t\t\t\tinstall\nfindutils\t\t\t\t\tinstall\nfontconfig\t\t\t\t\tinstall\nfontconfig-config\t\t\t\tinstall\nfonts-ubuntu-font-family-console\t\tinstall\nfriendly-recovery\t\t\t\tinstall\nftp\t\t\t\t\t\tinstall\nfuse\t\t\t\t\t\tinstall\ngcc-4.6-base\t\t\t\t\tinstall\ngconf-service\t\t\t\t\tinstall\ngconf-service-backend\t\t\t\tinstall\ngconf2\t\t\t\t\t\tinstall\ngconf2-common\t\t\t\t\tinstall\ngeoip-database\t\t\t\t\tinstall\ngettext-base\t\t\t\t\tinstall\ngir1.2-glib-2.0\t\t\t\t\tinstall\ngnupg\t\t\t\t\t\tinstall\ngpgv\t\t\t\t\t\tinstall\ngrep\t\t\t\t\t\tinstall\ngroff-base\t\t\t\t\tinstall\ngrub-common\t\t\t\t\tinstall\ngrub-gfxpayload-lists\t\t\t\tinstall\ngrub-legacy-ec2\t\t\t\t\tinstall\ngrub-pc\t\t\t\t\t\tinstall\ngrub-pc-bin\t\t\t\t\tinstall\ngrub2-common\t\t\t\t\tinstall\ngvfs\t\t\t\t\t\tinstall\ngvfs-common\t\t\t\t\tinstall\ngvfs-daemons\t\t\t\t\tinstall\ngvfs-libs\t\t\t\t\tinstall\ngzip\t\t\t\t\t\tinstall\nhdparm\t\t\t\t\t\tinstall\nhicolor-icon-theme\t\t\t\tinstall\nhostname\t\t\t\t\tinstall\nicedtea-7-jre-jamvm\t\t\t\tinstall\nifupdown\t\t\t\t\tinstall\ninfo\t\t\t\t\t\tinstall\ninitramfs-tools\t\t\t\t\tinstall\ninitramfs-tools-bin\t\t\t\tinstall\ninitscripts\t\t\t\t\tinstall\ninsserv\t\t\t\t\t\tinstall\ninstall-info\t\t\t\t\tinstall\niproute\t\t\t\t\t\tinstall\niptables\t\t\t\t\tinstall\niputils-ping\t\t\t\t\tinstall\niputils-tracepath\t\t\t\tinstall\nirqbalance\t\t\t\t\tinstall\nisc-dhcp-client\t\t\t\t\tinstall\nisc-dhcp-common\t\t\t\t\tinstall\niso-codes\t\t\t\t\tinstall\niw\t\t\t\t\t\tinstall\njava-common\t\t\t\t\tinstall\nkbd\t\t\t\t\t\tinstall\nkeyboard-configuration\t\t\t\tinstall\nklibc-utils\t\t\t\t\tinstall\nkrb5-locales\t\t\t\t\tinstall\nlandscape-client\t\t\t\tinstall\nlandscape-common\t\t\t\tinstall\nlanguage-selector-common\t\t\tinstall\nlaptop-detect\t\t\t\t\tinstall\nless\t\t\t\t\t\tinstall\nlibaccountsservice0\t\t\t\tinstall\nlibacl1\t\t\t\t\t\tinstall\nlibapt-inst1.4\t\t\t\t\tinstall\nlibapt-pkg4.12\t\t\t\t\tinstall\nlibasn1-8-heimdal\t\t\t\tinstall\nlibasound2\t\t\t\t\tinstall\nlibasyncns0\t\t\t\t\tinstall\nlibatasmart4\t\t\t\t\tinstall\nlibatk-wrapper-java\t\t\t\tinstall\nlibatk-wrapper-java-jni\t\t\t\tinstall\nlibatk1.0-0\t\t\t\t\tinstall\nlibatk1.0-data\t\t\t\t\tinstall\nlibattr1\t\t\t\t\tinstall\nlibavahi-client3\t\t\t\tinstall\nlibavahi-common-data\t\t\t\tinstall\nlibavahi-common3\t\t\t\tinstall\nlibavahi-glib1\t\t\t\t\tinstall\nlibbind9-80\t\t\t\t\tinstall\nlibblkid1\t\t\t\t\tinstall\nlibbonobo2-0\t\t\t\t\tinstall\nlibbonobo2-common\t\t\t\tinstall\nlibboost-iostreams1.46.1\t\t\tinstall\nlibbsd0\t\t\t\t\t\tinstall\nlibbz2-1.0\t\t\t\t\tinstall\nlibc-bin\t\t\t\t\tinstall\nlibc6\t\t\t\t\t\tinstall\nlibcairo-gobject2\t\t\t\tinstall\nlibcairo2\t\t\t\t\tinstall\nlibcanberra0\t\t\t\t\tinstall\nlibcap-ng0\t\t\t\t\tinstall\nlibck-connector0\t\t\t\tinstall\nlibclass-accessor-perl\t\t\t\tinstall\nlibclass-isa-perl\t\t\t\tinstall\nlibcomerr2\t\t\t\t\tinstall\nlibcryptsetup4\t\t\t\t\tinstall\nlibcups2\t\t\t\t\tinstall\nlibcurl3\t\t\t\t\tinstall\nlibcurl3-gnutls\t\t\t\t\tinstall\nlibcwidget3\t\t\t\t\tinstall\nlibdatrie1\t\t\t\t\tinstall\nlibdb5.1\t\t\t\t\tinstall\nlibdbus-1-3\t\t\t\t\tinstall\nlibdbus-glib-1-2\t\t\t\tinstall\nlibdconf0\t\t\t\t\tinstall\nlibdevmapper-event1.02.1\t\t\tinstall\nlibdevmapper1.02.1\t\t\t\tinstall\nlibdns81\t\t\t\t\tinstall\nlibdrm-intel1\t\t\t\t\tinstall\nlibdrm-nouveau1a\t\t\t\tinstall\nlibdrm-radeon1\t\t\t\t\tinstall\nlibdrm2\t\t\t\t\t\tinstall\nlibedit2\t\t\t\t\tinstall\nlibelf1\t\t\t\t\t\tinstall\nlibept1.4.12\t\t\t\t\tinstall\nlibevent-2.0-5\t\t\t\t\tinstall\nlibexpat1\t\t\t\t\tinstall\nlibffi6\t\t\t\t\t\tinstall\nlibflac8\t\t\t\t\tinstall\nlibfontconfig1\t\t\t\t\tinstall\nlibfontenc1\t\t\t\t\tinstall\nlibfreetype6\t\t\t\t\tinstall\nlibfribidi0\t\t\t\t\tinstall\nlibfuse2\t\t\t\t\tinstall\nlibgc1c2\t\t\t\t\tinstall\nlibgcc1\t\t\t\t\t\tinstall\nlibgconf-2-4\t\t\t\t\tinstall\nlibgconf2-4\t\t\t\t\tinstall\nlibgcrypt11\t\t\t\t\tinstall\nlibgdbm3\t\t\t\t\tinstall\nlibgdk-pixbuf2.0-0\t\t\t\tinstall\nlibgdk-pixbuf2.0-common\t\t\t\tinstall\nlibgdu0\t\t\t\t\t\tinstall\nlibgeoip1\t\t\t\t\tinstall\nlibgif4\t\t\t\t\t\tinstall\nlibgirepository-1.0-1\t\t\t\tinstall\nlibgl1-mesa-dri\t\t\t\t\tinstall\nlibgl1-mesa-glx\t\t\t\t\tinstall\nlibglapi-mesa\t\t\t\t\tinstall\nlibglib2.0-0\t\t\t\t\tinstall\nlibgmp10\t\t\t\t\tinstall\nlibgnome-keyring-common\t\t\t\tinstall\nlibgnome-keyring0\t\t\t\tinstall\nlibgnome2-0\t\t\t\t\tinstall\nlibgnome2-bin\t\t\t\t\tinstall\nlibgnome2-common\t\t\t\tinstall\nlibgnomevfs2-0\t\t\t\t\tinstall\nlibgnomevfs2-common\t\t\t\tinstall\nlibgnutls26\t\t\t\t\tinstall\nlibgpg-error0\t\t\t\t\tinstall\nlibgpm2\t\t\t\t\t\tinstall\nlibgssapi-krb5-2\t\t\t\tinstall\nlibgssapi3-heimdal\t\t\t\tinstall\nlibgtk-3-0\t\t\t\t\tinstall\nlibgtk-3-bin\t\t\t\t\tinstall\nlibgtk-3-common\t\t\t\t\tinstall\nlibgtk2.0-0\t\t\t\t\tinstall\nlibgtk2.0-bin\t\t\t\t\tinstall\nlibgtk2.0-common\t\t\t\tinstall\nlibgudev-1.0-0\t\t\t\t\tinstall\nlibhcrypto4-heimdal\t\t\t\tinstall\nlibheimbase1-heimdal\t\t\t\tinstall\nlibheimntlm0-heimdal\t\t\t\tinstall\nlibhx509-5-heimdal\t\t\t\tinstall\nlibice-dev\t\t\t\t\tinstall\nlibice6\t\t\t\t\t\tinstall\nlibidl-common\t\t\t\t\tinstall\nlibidl0\t\t\t\t\t\tinstall\nlibidn11\t\t\t\t\tinstall\nlibio-string-perl\t\t\t\tinstall\nlibisc83\t\t\t\t\tinstall\nlibisccc80\t\t\t\t\tinstall\nlibisccfg82\t\t\t\t\tinstall\nlibiw30\t\t\t\t\t\tinstall\nlibjasper1\t\t\t\t\tinstall\nlibjpeg-turbo8\t\t\t\t\tinstall\nlibjpeg8\t\t\t\t\tinstall\nlibjs-jquery\t\t\t\t\tinstall\nlibjson0\t\t\t\t\tinstall\nlibk5crypto3\t\t\t\t\tinstall\nlibkeyutils1\t\t\t\t\tinstall\nlibklibc\t\t\t\t\tinstall\nlibkrb5-26-heimdal\t\t\t\tinstall\nlibkrb5-3\t\t\t\t\tinstall\nlibkrb5support0\t\t\t\t\tinstall\nliblcms2-2\t\t\t\t\tinstall\nlibldap-2.4-2\t\t\t\t\tinstall\nlibllvm3.0\t\t\t\t\tinstall\nliblocale-gettext-perl\t\t\t\tinstall\nliblockfile-bin\t\t\t\t\tinstall\nliblockfile1\t\t\t\t\tinstall\nlibltdl7\t\t\t\t\tinstall\nliblvm2app2.2\t\t\t\t\tinstall\nliblwres80\t\t\t\t\tinstall\nliblzma5\t\t\t\t\tinstall\nlibmagic1\t\t\t\t\tinstall\nlibmount1\t\t\t\t\tinstall\nlibmpc2\t\t\t\t\t\tinstall\nlibmpfr4\t\t\t\t\tinstall\nlibncurses5\t\t\t\t\tinstall\nlibncursesw5\t\t\t\t\tinstall\nlibnewt0.52\t\t\t\t\tinstall\nlibnfnetlink0\t\t\t\t\tinstall\nlibnih-dbus1\t\t\t\t\tinstall\nlibnih1\t\t\t\t\t\tinstall\nlibnl-3-200\t\t\t\t\tinstall\nlibnl-genl-3-200\t\t\t\tinstall\nlibnspr4\t\t\t\t\tinstall\nlibnss3\t\t\t\t\t\tinstall\nlibnss3-1d\t\t\t\t\tinstall\nlibogg0\t\t\t\t\t\tinstall\nliborbit2\t\t\t\t\tinstall\nlibp11-kit0\t\t\t\t\tinstall\nlibpam-ck-connector\t\t\t\tinstall\nlibpam-modules\t\t\t\t\tinstall\nlibpam-modules-bin\t\t\t\tinstall\nlibpam-runtime\t\t\t\t\tinstall\nlibpam0g\t\t\t\t\tinstall\nlibpango1.0-0\t\t\t\t\tinstall\nlibparse-debianchangelog-perl\t\t\tinstall\nlibparted0debian1\t\t\t\tinstall\nlibpcap0.8\t\t\t\t\tinstall\nlibpci3\t\t\t\t\t\tinstall\nlibpciaccess0\t\t\t\t\tinstall\nlibpcre3\t\t\t\t\tinstall\nlibpcsclite1\t\t\t\t\tinstall\nlibpipeline1\t\t\t\t\tinstall\nlibpixman-1-0\t\t\t\t\tinstall\nlibplymouth2\t\t\t\t\tinstall\nlibpng12-0\t\t\t\t\tinstall\nlibpolkit-agent-1-0\t\t\t\tinstall\nlibpolkit-backend-1-0\t\t\t\tinstall\nlibpolkit-gobject-1-0\t\t\t\tinstall\nlibpopt0\t\t\t\t\tinstall\nlibpthread-stubs0\t\t\t\tinstall\nlibpthread-stubs0-dev\t\t\t\tinstall\nlibpulse0\t\t\t\t\tinstall\nlibpython2.7\t\t\t\t\tinstall\nlibreadline6\t\t\t\t\tinstall\nlibroken18-heimdal\t\t\t\tinstall\nlibrtmp0\t\t\t\t\tinstall\nlibsasl2-2\t\t\t\t\tinstall\nlibsasl2-modules\t\t\t\tinstall\nlibselinux1\t\t\t\t\tinstall\nlibsgutils2-2\t\t\t\t\tinstall\nlibsigc++-2.0-0c2a\t\t\t\tinstall\nlibslang2\t\t\t\t\tinstall\nlibsm-dev\t\t\t\t\tinstall\nlibsm6\t\t\t\t\t\tinstall\nlibsndfile1\t\t\t\t\tinstall\nlibsqlite3-0\t\t\t\t\tinstall\nlibss2\t\t\t\t\t\tinstall\nlibssl1.0.0\t\t\t\t\tinstall\nlibstdc++6\t\t\t\t\tinstall\nlibsub-name-perl\t\t\t\tinstall\nlibswitch-perl\t\t\t\t\tinstall\nlibtasn1-3\t\t\t\t\tinstall\nlibtdb1\t\t\t\t\t\tinstall\nlibtext-charwidth-perl\t\t\t\tinstall\nlibtext-iconv-perl\t\t\t\tinstall\nlibtext-wrapi18n-perl\t\t\t\tinstall\nlibthai-data\t\t\t\t\tinstall\nlibthai0\t\t\t\t\tinstall\nlibtiff4\t\t\t\t\tinstall\nlibtimedate-perl\t\t\t\tinstall\nlibtinfo5\t\t\t\t\tinstall\nlibudev0\t\t\t\t\tinstall\nlibusb-0.1-4\t\t\t\t\tinstall\nlibusb-1.0-0\t\t\t\t\tinstall\nlibuuid1\t\t\t\t\tinstall\nlibvorbis0a\t\t\t\t\tinstall\nlibvorbisenc2\t\t\t\t\tinstall\nlibvorbisfile3\t\t\t\t\tinstall\nlibwind0-heimdal\t\t\t\tinstall\nlibwrap0\t\t\t\t\tinstall\nlibx11-6\t\t\t\t\tinstall\nlibx11-data\t\t\t\t\tinstall\nlibx11-dev\t\t\t\t\tinstall\nlibx11-doc\t\t\t\t\tinstall\nlibx11-xcb1\t\t\t\t\tinstall\nlibxapian22\t\t\t\t\tinstall\nlibxau-dev\t\t\t\t\tinstall\nlibxau6\t\t\t\t\t\tinstall\nlibxaw7\t\t\t\t\t\tinstall\nlibxcb-glx0\t\t\t\t\tinstall\nlibxcb-render0\t\t\t\t\tinstall\nlibxcb-shape0\t\t\t\t\tinstall\nlibxcb-shm0\t\t\t\t\tinstall\nlibxcb1\t\t\t\t\t\tinstall\nlibxcb1-dev\t\t\t\t\tinstall\nlibxcomposite1\t\t\t\t\tinstall\nlibxcursor1\t\t\t\t\tinstall\nlibxdamage1\t\t\t\t\tinstall\nlibxdmcp-dev\t\t\t\t\tinstall\nlibxdmcp6\t\t\t\t\tinstall\nlibxext6\t\t\t\t\tinstall\nlibxfixes3\t\t\t\t\tinstall\nlibxft2\t\t\t\t\t\tinstall\nlibxi6\t\t\t\t\t\tinstall\nlibxinerama1\t\t\t\t\tinstall\nlibxml2\t\t\t\t\t\tinstall\nlibxmu6\t\t\t\t\t\tinstall\nlibxmuu1\t\t\t\t\tinstall\nlibxpm4\t\t\t\t\t\tinstall\nlibxrandr2\t\t\t\t\tinstall\nlibxrender1\t\t\t\t\tinstall\nlibxt-dev\t\t\t\t\tinstall\nlibxt6\t\t\t\t\t\tinstall\nlibxtst6\t\t\t\t\tinstall\nlibxv1\t\t\t\t\t\tinstall\nlibxxf86dga1\t\t\t\t\tinstall\nlibxxf86vm1\t\t\t\t\tinstall\nlibyaml-0-2\t\t\t\t\tinstall\nlinux-firmware\t\t\t\t\tinstall\nlinux-headers-3.2.0-60\t\t\t\tinstall\nlinux-headers-3.2.0-60-virtual\t\t\tinstall\nlinux-headers-virtual\t\t\t\tinstall\nlinux-image-3.2.0-60-virtual\t\t\tinstall\nlinux-image-virtual\t\t\t\tinstall\nlinux-virtual\t\t\t\t\tinstall\nlocales\t\t\t\t\t\tinstall\nlockfile-progs\t\t\t\t\tinstall\nlogin\t\t\t\t\t\tinstall\nlogrotate\t\t\t\t\tinstall\nlsb-base\t\t\t\t\tinstall\nlsb-release\t\t\t\t\tinstall\nlshw\t\t\t\t\t\tinstall\nlsof\t\t\t\t\t\tinstall\nltrace\t\t\t\t\t\tinstall\nmakedev\t\t\t\t\t\tinstall\nman-db\t\t\t\t\t\tinstall\nmanpages\t\t\t\t\tinstall\nmawk\t\t\t\t\t\tinstall\nmemtest86+\t\t\t\t\tinstall\nmime-support\t\t\t\t\tinstall\nmlocate\t\t\t\t\t\tinstall\nmodule-init-tools\t\t\t\tinstall\nmount\t\t\t\t\t\tinstall\nmountall\t\t\t\t\tinstall\nmtools\t\t\t\t\t\tinstall\nmtr-tiny\t\t\t\t\tinstall\nmultiarch-support\t\t\t\tinstall\nnano\t\t\t\t\t\tinstall\nncurses-base\t\t\t\t\tinstall\nncurses-bin\t\t\t\t\tinstall\nnet-tools\t\t\t\t\tinstall\nnetbase\t\t\t\t\t\tinstall\nnetcat-openbsd\t\t\t\t\tinstall\nntfs-3g\t\t\t\t\t\tinstall\nntpdate\t\t\t\t\t\tinstall\nopenjdk-7-jdk\t\t\t\t\tinstall\nopenjdk-7-jre\t\t\t\t\tinstall\nopenjdk-7-jre-headless\t\t\t\tinstall\nopenssh-client\t\t\t\t\tinstall\nopenssh-server\t\t\t\t\tinstall\nopenssl\t\t\t\t\t\tinstall\nos-prober\t\t\t\t\tinstall\nparted\t\t\t\t\t\tinstall\npasswd\t\t\t\t\t\tinstall\npatch\t\t\t\t\t\tinstall\npciutils\t\t\t\t\tinstall\nperl\t\t\t\t\t\tinstall\nperl-base\t\t\t\t\tinstall\nperl-modules\t\t\t\t\tinstall\nplymouth\t\t\t\t\tinstall\nplymouth-theme-ubuntu-text\t\t\tinstall\npolicykit-1\t\t\t\t\tinstall\npolicykit-1-gnome\t\t\t\tinstall\npopularity-contest\t\t\t\tinstall\npowermgmt-base\t\t\t\t\tinstall\nppp\t\t\t\t\t\tinstall\npppconfig\t\t\t\t\tinstall\npppoeconf\t\t\t\t\tinstall\nprocps\t\t\t\t\t\tinstall\npsmisc\t\t\t\t\t\tinstall\npython\t\t\t\t\t\tinstall\npython-apport\t\t\t\t\tinstall\npython-apt\t\t\t\t\tinstall\npython-apt-common\t\t\t\tinstall\npython-boto\t\t\t\t\tinstall\npython-chardet\t\t\t\t\tinstall\npython-cheetah\t\t\t\t\tinstall\npython-configobj\t\t\t\tinstall\npython-crypto\t\t\t\t\tinstall\npython-dbus\t\t\t\t\tinstall\npython-dbus-dev\t\t\t\t\tinstall\npython-debian\t\t\t\t\tinstall\npython-gdbm\t\t\t\t\tinstall\npython-gi\t\t\t\t\tinstall\npython-gnupginterface\t\t\t\tinstall\npython-httplib2\t\t\t\t\tinstall\npython-keyring\t\t\t\t\tinstall\npython-launchpadlib\t\t\t\tinstall\npython-lazr.restfulclient\t\t\tinstall\npython-lazr.uri\t\t\t\t\tinstall\npython-m2crypto\t\t\t\t\tinstall\npython-minimal\t\t\t\t\tinstall\npython-newt\t\t\t\t\tinstall\npython-oauth\t\t\t\t\tinstall\npython-openssl\t\t\t\t\tinstall\npython-pam\t\t\t\t\tinstall\npython-paramiko\t\t\t\t\tinstall\npython-pkg-resources\t\t\t\tinstall\npython-problem-report\t\t\t\tinstall\npython-pycurl\t\t\t\t\tinstall\npython-serial\t\t\t\t\tinstall\npython-simplejson\t\t\t\tinstall\npython-software-properties\t\t\tinstall\npython-twisted-bin\t\t\t\tinstall\npython-twisted-core\t\t\t\tinstall\npython-twisted-names\t\t\t\tinstall\npython-twisted-web\t\t\t\tinstall\npython-wadllib\t\t\t\t\tinstall\npython-xapian\t\t\t\t\tinstall\npython-yaml\t\t\t\t\tinstall\npython-zope.interface\t\t\t\tinstall\npython2.7\t\t\t\t\tinstall\npython2.7-minimal\t\t\t\tinstall\nreadline-common\t\t\t\t\tinstall\nresolvconf\t\t\t\t\tinstall\nrsync\t\t\t\t\t\tinstall\nrsyslog\t\t\t\t\t\tinstall\nscreen\t\t\t\t\t\tinstall\nsed\t\t\t\t\t\tinstall\nsensible-utils\t\t\t\t\tinstall\nsgml-base\t\t\t\t\tinstall\nshared-mime-info\t\t\t\tinstall\nsound-theme-freedesktop\t\t\t\tinstall\nssh-import-id\t\t\t\t\tinstall\nstrace\t\t\t\t\t\tinstall\nsudo\t\t\t\t\t\tinstall\nsysv-rc\t\t\t\t\t\tinstall\nsysvinit-utils\t\t\t\t\tinstall\ntar\t\t\t\t\t\tinstall\ntasksel\t\t\t\t\t\tinstall\ntasksel-data\t\t\t\t\tinstall\ntcpd\t\t\t\t\t\tinstall\ntcpdump\t\t\t\t\t\tinstall\ntelnet\t\t\t\t\t\tinstall\ntime\t\t\t\t\t\tinstall\ntmux\t\t\t\t\t\tinstall\nttf-dejavu-core\t\t\t\t\tinstall\nttf-dejavu-extra\t\t\t\tinstall\ntzdata\t\t\t\t\t\tinstall\ntzdata-java\t\t\t\t\tinstall\nubuntu-keyring\t\t\t\t\tinstall\nubuntu-minimal\t\t\t\t\tinstall\nubuntu-standard\t\t\t\t\tinstall\nucf\t\t\t\t\t\tinstall\nudev\t\t\t\t\t\tinstall\nudisks\t\t\t\t\t\tinstall\nufw\t\t\t\t\t\tinstall\nunattended-upgrades\t\t\t\tinstall\nupdate-manager-core\t\t\t\tinstall\nupdate-notifier-common\t\t\t\tinstall\nupstart\t\t\t\t\t\tinstall\nureadahead\t\t\t\t\tinstall\nusbutils\t\t\t\t\tinstall\nutil-linux\t\t\t\t\tinstall\nuuid-runtime\t\t\t\t\tinstall\nvim\t\t\t\t\t\tinstall\nvim-common\t\t\t\t\tinstall\nvim-runtime\t\t\t\t\tinstall\nvim-tiny\t\t\t\t\tinstall\nw3m\t\t\t\t\t\tinstall\nwget\t\t\t\t\t\tinstall\nwhiptail\t\t\t\t\tinstall\nwhoopsie\t\t\t\t\tinstall\nwireless-regdb\t\t\t\t\tinstall\nwireless-tools\t\t\t\t\tinstall\nwpasupplicant\t\t\t\t\tinstall\nx11-common\t\t\t\t\tinstall\nx11-utils\t\t\t\t\tinstall\nx11proto-core-dev\t\t\t\tinstall\nx11proto-input-dev\t\t\t\tinstall\nx11proto-kb-dev\t\t\t\t\tinstall\nxauth\t\t\t\t\t\tinstall\nxkb-data\t\t\t\t\tinstall\nxml-core\t\t\t\t\tinstall\nxorg-sgml-doctools\t\t\t\tinstall\nxtrans-dev\t\t\t\t\tinstall\nxz-lzma\t\t\t\t\t\tinstall\nxz-utils\t\t\t\t\tinstall\nzlib1g\t\t\t\t\t\tinstall\n#> [install: install: install-from]: Packages : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho '[install: install: install-from]: Packages...';\n{\n{ debconf-set-selections <<EOF\ndebconf debconf/frontend select noninteractive\ndebconf debconf/frontend seen false\nEOF\n} && enableStart() {\nrm /usr/sbin/policy-rc.d\n} && apt-get -q -y install openjdk-7-jdk+ && dpkg --get-selections\n } || { echo '#> [install: install: install-from]: Packages : FAIL'; exit 1;} >&2 \necho '#> [install: install: install-from]: Packages : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package, | |
:context "[install: install: install-from]: "} | |
{:exit 0, | |
:out "", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent group hadoop; then /usr/sbin/groupmod hadoop;else /usr/sbin/groupadd --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/group, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out "hadoop:x:999:\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent group hadoop; then /usr/sbin/groupmod hadoop;else /usr/sbin/groupadd --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/group, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out "hadoop:x:999:999::/home/hadoop:/bin/bash\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent passwd hadoop; then /usr/sbin/usermod --group \"hadoop\" --shell \"/bin/bash\" hadoop;else /usr/sbin/useradd -g \"hadoop\" --shell \"/bin/bash\" --create-home --system hadoop;fi\nif getent passwd hadoop; then /usr/sbin/usermod --group \"hadoop\" --shell \"/bin/bash\" hadoop;else /usr/sbin/useradd -g \"hadoop\" --shell \"/bin/bash\" --create-home --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/user, | |
:context "[hadoop-user]: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install: set-environment: system-environment: plan-when: Add java environment to /etc/environment...';\n{\nif ! ( [ -e /etc/environment ] ); then\n{ cat > /etc/environment <<EOFpallet\n# environment file created by pallet\n\nEOFpallet\n }\nfi\npallet_set_env() {\nk=$1\nv=$2\ns=$3\nif ! ( grep \"${s}\" /etc/environment 2>&- ); then\nsed -i -e \"/$${k}=/ d\" /etc/environment && sed -i -e \"$ a \\\\\n${s}\" /etc/environment || exit 1\nfi\n} && vv=\"$(dirname $(dirname $(update-alternatives --query javac | grep Best: | cut -f 2 -d ' ')))\"\npallet_set_env \"JAVA_HOME\" \"${vv}\" \"JAVA_HOME=\\\"${vv}\\\"\"\n } || { echo '#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : FAIL'; exit 1;} >&2 \necho '#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"install: set-environment: system-environment: plan-when: Add java environment to /etc/environment...\n#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"install: set-environment: system-environment: plan-when: "} | |
nil | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/dfs/data && chown hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/data && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/data\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/dfs/name && chown hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/name && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/name\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)//dfs/secondary && chown hadoop $(getent passwd hadoop | cut -d: -f6)//dfs/secondary && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)//dfs/secondary\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/mapred/local && chown hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/local && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/local\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/mapred/system && chown hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/system && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/system\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory /var/log/hadoop...\n#> local-dirs: Directory /var/log/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory /var/log/hadoop...';\n{\nmkdir -p /var/log/hadoop && chown hadoop /var/log/hadoop && chgrp hadoop /var/log/hadoop\n } || { echo '#> local-dirs: Directory /var/log/hadoop : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory /var/log/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory /var/run/hadoop...\n#> local-dirs: Directory /var/run/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory /var/run/hadoop...';\n{\nmkdir -p /var/run/hadoop && chown hadoop /var/run/hadoop && chgrp hadoop /var/run/hadoop\n } || { echo '#> local-dirs: Directory /var/run/hadoop : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory /var/run/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...\nMD5 sum is 2402b944f6cd21bed1003d9e13107221 .bash_profile\n#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) <<'EOFpallet'\n # base.clj:351\nJAVA_HOME=$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\n # base.clj:352\nPATH=${PATH}:/usr/local/hadoop-0.20.2/bin:/usr/local/hadoop-0.20.2/bin\n\nEOFpallet\n } && contentdiff= && if [ -e $(getent passwd user | cut -d: -f6)/.bash_profile ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\ndiff -u $(getent passwd user | cut -d: -f6)/.bash_profile ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) $(getent passwd user | cut -d: -f6)/.bash_profile && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n}\nfi\n } || { echo '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : FAIL'; exit 1;} >&2 \necho '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out | |
"create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...\n#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...';\n{\nln -f -s /usr/local/hadoop-0.20.2/conf /etc/hadoop\n } || { echo '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : FAIL'; exit 1;} >&2 \necho '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/symbolic-link, | |
:context "create-directories: "} | |
{:exit 0, | |
:out | |
"install-hadoop: install: remote-directory...\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\ninstall-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...\n#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS\ninstall-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\nMD5 sum is 1ddf6e8ff716d17631a46384544293c1 hadoop-0.20.2-cdh3u5.tar.gz\n#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\nUntar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n/home/tom\n#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\n#> install-hadoop: install: remote-directory : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-hadoop: install: remote-directory...';\n{\necho 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown hadoop /usr/local/hadoop-0.20.2 && chgrp hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS' && echo 'install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...';\n{\nmkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\n } || { echo '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS' && echo 'install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))) && if hash curl 2>&-; then\ncurl -o \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --retry 5 --silent --show-error --fail --location \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\nif hash wget 2>&-; then\nwget -O \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --tries 5 --no-verbose --progress=dot:mega \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\necho No download utility available\nexit 1\nfi\nfi && contentdiff= && if [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\ndiff -u ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\nmv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n}\nfi\n } || { echo '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS' && if ! { [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; } || ! { [ -e /usr/local/hadoop-0.20.2/.pallet.directory.md5 ]; } || ! { diff ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5; }; then\necho 'Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\nrdf=$(readlink -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) && cd /usr/local/hadoop-0.20.2 && tar xz --strip-components=1 -f ${rdf} && cd -\n } || { echo '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS'\nif [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; then\ncp ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5\nfi\nfi && echo 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown --recursive hadoop /usr/local/hadoop-0.20.2 && chgrp --recursive hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS'\n } || { echo '#> install-hadoop: install: remote-directory : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-directory : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-directory-action, | |
:context "install-hadoop: install: "} | |
{:exit 0, | |
:out | |
"hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...\n#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) <<'EOFpallet'\n # base.clj:351\nJAVA_HOME=$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\n # base.clj:352\nPATH=${PATH}:/usr/local/hadoop-0.20.2/bin:/usr/local/hadoop-0.20.2/bin\n\nEOFpallet\n } && contentdiff= && if [ -e $(getent passwd user | cut -d: -f6)/.bash_profile ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\ndiff -u $(getent passwd user | cut -d: -f6)/.bash_profile ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) $(getent passwd user | cut -d: -f6)/.bash_profile && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n}\nfi\n } || { echo '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : FAIL'; exit 1;} >&2 \necho '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out | |
"create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...\n#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...';\n{\nln -f -s /usr/local/hadoop-0.20.2/conf /etc/hadoop\n } || { echo '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : FAIL'; exit 1;} >&2 \necho '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/symbolic-link, | |
:context "create-directories: "} | |
{:exit 0, | |
:out | |
"install-hadoop: install: remote-directory...\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\ninstall-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...\n#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS\ninstall-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\nUntar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n/home/tom\n#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\n#> install-hadoop: install: remote-directory : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-hadoop: install: remote-directory...';\n{\necho 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown hadoop /usr/local/hadoop-0.20.2 && chgrp hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS' && echo 'install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...';\n{\nmkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\n } || { echo '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS' && echo 'install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))) && if hash curl 2>&-; then\ncurl -o \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --retry 5 --silent --show-error --fail --location \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\nif hash wget 2>&-; then\nwget -O \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --tries 5 --no-verbose --progress=dot:mega \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\necho No download utility available\nexit 1\nfi\nfi && contentdiff= && if [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\ndiff -u ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\nmv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n}\nfi\n } || { echo '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS' && if ! { [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; } || ! { [ -e /usr/local/hadoop-0.20.2/.pallet.directory.md5 ]; } || ! { diff ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5; }; then\necho 'Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\nrdf=$(readlink -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) && cd /usr/local/hadoop-0.20.2 && tar xz --strip-components=1 -f ${rdf} && cd -\n } || { echo '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS'\nif [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; then\ncp ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5\nfi\nfi && echo 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown --recursive hadoop /usr/local/hadoop-0.20.2 && chgrp --recursive hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS'\n } || { echo '#> install-hadoop: install: remote-directory : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-directory : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-directory-action, | |
:context "install-hadoop: install: "} | |
{:exit 0, | |
:out | |
"install-ssh-key: generate-key: Directory $(getent passwd hadoop | cut -d: -f6)/.ssh...\n#> install-ssh-key: generate-key: Directory $(getent passwd hadoop | cut -d: -f6)/.ssh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-ssh-key: generate-key: Directory $(getent passwd hadoop | cut -d: -f6)/.ssh...';\n{\nmkdir -m \"755\" -p $(getent passwd hadoop | cut -d: -f6)/.ssh && chown --recursive hadoop $(getent passwd hadoop | cut -d: -f6)/.ssh && chmod 755 $(getent passwd hadoop | cut -d: -f6)/.ssh\n } || { echo '#> install-ssh-key: generate-key: Directory $(getent passwd hadoop | cut -d: -f6)/.ssh : FAIL'; exit 1;} >&2 \necho '#> install-ssh-key: generate-key: Directory $(getent passwd hadoop | cut -d: -f6)/.ssh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "install-ssh-key: generate-key: "} | |
{:exit 0, | |
:out | |
"install-ssh-key: generate-key: ssh-keygen (ssh_key.clj:99)...\nGenerating public/private rsa key pair.\nYour identification has been saved in /home/hadoop/.ssh/id_rsa.\nYour public key has been saved in /home/hadoop/.ssh/id_rsa.pub.\nThe key fingerprint is:\nf2:01:77:57:9a:25:f0:0d:5e:80:f1:98:5a:e3:4b:0f ph_:hi-hadoop-nn_us-east-1/i-db664188_key\nThe key's randomart image is:\n+--[ RSA 2048]----+\n| o++.+ |\n| .* X |\n| . . * B . |\n| o = o |\n| . S E |\n| o o + |\n| . . . |\n| |\n| |\n+-----------------+\n#> install-ssh-key: generate-key: ssh-keygen (ssh_key.clj:99) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-ssh-key: generate-key: ssh-keygen (ssh_key.clj:99)...';\n{\nkey_path=$(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa && if ! ( [ -e ${key_path} ] ); then ssh-keygen -f \"${key_path}\" -t \"rsa\" -N \"\" -C \"ph_:hi-hadoop-nn_us-east-1/i-db664188_key\";fi\n } || { echo '#> install-ssh-key: generate-key: ssh-keygen (ssh_key.clj:99) : FAIL'; exit 1;} >&2 \necho '#> install-ssh-key: generate-key: ssh-keygen (ssh_key.clj:99) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "install-ssh-key: generate-key: "} | |
{:exit 0, | |
:out | |
"install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa...\n#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa...';\n{\ntouch $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa && chown hadoop $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa && chmod 0600 $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa\n } || { echo '#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa : FAIL'; exit 1;} >&2 \necho '#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/file, | |
:context "install-ssh-key: generate-key: "} | |
{:exit 0, | |
:out | |
"install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub...\n#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub...';\n{\ntouch $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub && chown hadoop $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub && chmod 0644 $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub\n } || { echo '#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub : FAIL'; exit 1;} >&2 \necho '#> install-ssh-key: generate-key: file $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/file, | |
:context "install-ssh-key: generate-key: "}), | |
:phase :install} | |
{:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"[install: install: install-from]: Packages...\nReading package lists...\nBuilding dependency tree...\nReading state information...\nThe following extra packages will be installed:\n ca-certificates-java consolekit cpp cpp-4.6 cryptsetup-bin dbus-x11\n dconf-gsettings-backend dconf-service fontconfig fontconfig-config\n gconf-service gconf-service-backend gconf2 gconf2-common gvfs gvfs-common\n gvfs-daemons gvfs-libs hicolor-icon-theme icedtea-7-jre-jamvm java-common\n libasound2 libasyncns0 libatasmart4 libatk-wrapper-java\n libatk-wrapper-java-jni libatk1.0-0 libatk1.0-data libavahi-client3\n libavahi-common-data libavahi-common3 libavahi-glib1 libbonobo2-0\n libbonobo2-common libcairo-gobject2 libcairo2 libcanberra0 libck-connector0\n libcryptsetup4 libcups2 libdatrie1 libdconf0 libdevmapper-event1.02.1\n libflac8 libfontconfig1 libfontenc1 libgconf-2-4 libgconf2-4\n libgdk-pixbuf2.0-0 libgdk-pixbuf2.0-common libgdu0 libgif4 libgl1-mesa-dri\n libgl1-mesa-glx libglapi-mesa libgnome-keyring-common libgnome-keyring0\n libgnome2-0 libgnome2-bin libgnome2-common libgnomevfs2-0\n libgnomevfs2-common libgtk-3-0 libgtk-3-bin libgtk-3-common libgtk2.0-0\n libgtk2.0-bin libgtk2.0-common libgudev-1.0-0 libice-dev libice6\n libidl-common libidl0 libjasper1 libjpeg-turbo8 libjpeg8 libjson0 liblcms2-2\n libllvm3.0 libltdl7 liblvm2app2.2 libmpc2 libmpfr4 libnspr4 libnss3\n libnss3-1d libogg0 liborbit2 libpam-ck-connector libpango1.0-0 libpixman-1-0\n libpolkit-agent-1-0 libpolkit-backend-1-0 libpthread-stubs0\n libpthread-stubs0-dev libpulse0 libsgutils2-2 libsm-dev libsm6 libsndfile1\n libtdb1 libthai-data libthai0 libtiff4 libvorbis0a libvorbisenc2\n libvorbisfile3 libx11-dev libx11-doc libx11-xcb1 libxau-dev libxaw7\n libxcb-glx0 libxcb-render0 libxcb-shape0 libxcb-shm0 libxcb1-dev\n libxcomposite1 libxcursor1 libxdamage1 libxdmcp-dev libxfixes3 libxft2\n libxi6 libxinerama1 libxmu6 libxpm4 libxrandr2 libxrender1 libxt-dev libxt6\n libxtst6 libxv1 libxxf86dga1 libxxf86vm1 mtools openjdk-7-jre\n openjdk-7-jre-headless policykit-1 policykit-1-gnome shared-mime-info\n sound-theme-freedesktop ttf-dejavu-core ttf-dejavu-extra tzdata-java udisks\n x11-common x11-utils x11proto-core-dev x11proto-input-dev x11proto-kb-dev\n xorg-sgml-doctools xtrans-dev\nSuggested packages:\n cpp-doc gcc-4.6-locales gconf-defaults-service gvfs-backends default-jre\n equivs libasound2-plugins libasound2-python libbonobo2-bin libcanberra-gtk0\n libcanberra-pulse cups-common libglide3 gnome-keyring desktop-base\n libgnomevfs2-bin libgnomevfs2-extra gamin fam gnome-mime-data\n librsvg2-common libjasper-runtime liblcms2-utils ttf-baekmuk\n ttf-arphic-gbsn00lp ttf-arphic-bsmi00lp ttf-arphic-gkai00mp\n ttf-arphic-bkai00mp pulseaudio sg3-utils libxcb-doc floppyd openjdk-7-demo\n openjdk-7-source visualvm icedtea-7-plugin libnss-mdns sun-java6-fonts\n fonts-ipafont-gothic fonts-ipafont-mincho ttf-wqy-microhei ttf-wqy-zenhei\n ttf-indic-fonts-core ttf-telugu-fonts ttf-oriya-fonts ttf-kannada-fonts\n ttf-bengali-fonts xfsprogs reiserfsprogs mdadm mesa-utils\nThe following NEW packages will be installed:\n ca-certificates-java consolekit cpp cpp-4.6 cryptsetup-bin dbus-x11\n dconf-gsettings-backend dconf-service fontconfig fontconfig-config\n gconf-service gconf-service-backend gconf2 gconf2-common gvfs gvfs-common\n gvfs-daemons gvfs-libs hicolor-icon-theme icedtea-7-jre-jamvm java-common\n libasound2 libasyncns0 libatasmart4 libatk-wrapper-java\n libatk-wrapper-java-jni libatk1.0-0 libatk1.0-data libavahi-client3\n libavahi-common-data libavahi-common3 libavahi-glib1 libbonobo2-0\n libbonobo2-common libcairo-gobject2 libcairo2 libcanberra0 libck-connector0\n libcryptsetup4 libcups2 libdatrie1 libdconf0 libdevmapper-event1.02.1\n libflac8 libfontconfig1 libfontenc1 libgconf-2-4 libgconf2-4\n libgdk-pixbuf2.0-0 libgdk-pixbuf2.0-common libgdu0 libgif4 libgl1-mesa-dri\n libgl1-mesa-glx libglapi-mesa libgnome-keyring-common libgnome-keyring0\n libgnome2-0 libgnome2-bin libgnome2-common libgnomevfs2-0\n libgnomevfs2-common libgtk-3-0 libgtk-3-bin libgtk-3-common libgtk2.0-0\n libgtk2.0-bin libgtk2.0-common libgudev-1.0-0 libice-dev libice6\n libidl-common libidl0 libjasper1 libjpeg-turbo8 libjpeg8 libjson0 liblcms2-2\n libllvm3.0 libltdl7 liblvm2app2.2 libmpc2 libmpfr4 libnspr4 libnss3\n libnss3-1d libogg0 liborbit2 libpam-ck-connector libpango1.0-0 libpixman-1-0\n libpolkit-agent-1-0 libpolkit-backend-1-0 libpthread-stubs0\n libpthread-stubs0-dev libpulse0 libsgutils2-2 libsm-dev libsm6 libsndfile1\n libtdb1 libthai-data libthai0 libtiff4 libvorbis0a libvorbisenc2\n libvorbisfile3 libx11-dev libx11-doc libx11-xcb1 libxau-dev libxaw7\n libxcb-glx0 libxcb-render0 libxcb-shape0 libxcb-shm0 libxcb1-dev\n libxcomposite1 libxcursor1 libxdamage1 libxdmcp-dev libxfixes3 libxft2\n libxi6 libxinerama1 libxmu6 libxpm4 libxrandr2 libxrender1 libxt-dev libxt6\n libxtst6 libxv1 libxxf86dga1 libxxf86vm1 mtools openjdk-7-jdk openjdk-7-jre\n openjdk-7-jre-headless policykit-1 policykit-1-gnome shared-mime-info\n sound-theme-freedesktop ttf-dejavu-core ttf-dejavu-extra tzdata-java udisks\n x11-common x11-utils x11proto-core-dev x11proto-input-dev x11proto-kb-dev\n xorg-sgml-doctools xtrans-dev\n0 upgraded, 154 newly installed, 0 to remove and 8 not upgraded.\nNeed to get 100 MB of archives.\nAfter this operation, 220 MB of additional disk space will be used.\nGet:1 http://security.ubuntu.com/ubuntu/ precise-security/main libcups2 amd64 1.5.3-0ubuntu8.2 [171 kB]\nGet:2 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libasound2 amd64 1.0.25-1ubuntu10.2 [428 kB]\nGet:3 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libasyncns0 amd64 0.8-4 [13.2 kB]\nGet:4 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatasmart4 amd64 0.18-3 [26.9 kB]\nGet:5 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk1.0-data all 2.4.0-0ubuntu1 [12.7 kB]\nGet:6 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk1.0-0 amd64 2.4.0-0ubuntu1 [60.1 kB]\nGet:7 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-common all 2.24.10-0ubuntu6.1 [118 kB]\nGet:8 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ttf-dejavu-core all 2.33-2ubuntu1 [1,552 kB]\nGet:9 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main fontconfig-config all 2.8.0-3ubuntu9.1 [44.4 kB]\nGet:10 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libfontconfig1 amd64 2.8.0-3ubuntu9.1 [125 kB]\nGet:11 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpixman-1-0 amd64 0.30.2-1ubuntu0.0.0.0.1 [256 kB]\nGet:12 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-render0 amd64 1.8.1-1ubuntu0.2 [13.3 kB]\nGet:13 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-shm0 amd64 1.8.1-1ubuntu0.2 [5,516 B]\nGet:14 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxrender1 amd64 1:0.9.6-2ubuntu0.1 [20.6 kB]\nGet:15 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libcairo2 amd64 1.10.2-6.1ubuntu3 [467 kB]\nGet:16 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-common-data amd64 0.6.30-5ubuntu2.1 [21.4 kB]\nGet:17 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-common3 amd64 0.6.30-5ubuntu2.1 [25.5 kB]\nGet:18 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-client3 amd64 0.6.30-5ubuntu2.1 [30.3 kB]\nGet:19 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libjpeg-turbo8 amd64 1.1.90+svn733-0ubuntu4.3 [111 kB]\nGet:20 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjpeg8 amd64 8c-2ubuntu7 [2,112 B]\nGet:21 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjasper1 amd64 1.900.1-13 [154 kB]\nGet:22 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libtiff4 amd64 3.9.5-2ubuntu1.5 [145 kB]\nGet:23 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgdk-pixbuf2.0-common all 2.26.1-1ubuntu1.1 [7,846 B]\nGet:24 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgdk-pixbuf2.0-0 amd64 2.26.1-1ubuntu1.1 [200 kB]\nGet:25 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libthai-data all 0.1.16-3 [179 kB]\nGet:26 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libdatrie1 amd64 0.2.5-3 [16.4 kB]\nGet:27 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libthai0 amd64 0.1.16-3 [19.7 kB]\nGet:28 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxft2 amd64 2.2.0-3ubuntu2 [42.3 kB]\nGet:29 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main fontconfig amd64 2.8.0-3ubuntu9.1 [156 kB]\nGet:30 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpango1.0-0 amd64 1.30.0-0ubuntu3.1 [363 kB]\nGet:31 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxcomposite1 amd64 1:0.4.3-2build1 [7,708 B]\nGet:32 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxfixes3 amd64 1:5.0-4ubuntu4.2 [12.5 kB]\nGet:33 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcursor1 amd64 1:1.1.12-1ubuntu0.1 [22.7 kB]\nGet:34 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxdamage1 amd64 1:1.1.3-2build1 [7,560 B]\nGet:35 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxi6 amd64 2:1.7.1.901-1ubuntu1~precise1 [31.6 kB]\nGet:36 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxinerama1 amd64 2:1.1.1-3ubuntu0.1 [8,072 B]\nGet:37 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxrandr2 amd64 2:1.3.2-2ubuntu0.2 [17.5 kB]\nGet:38 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main shared-mime-info amd64 1.0-0ubuntu4.1 [496 kB]\nGet:39 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-0 amd64 2.24.10-0ubuntu6.1 [2,674 kB]\nGet:40 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnspr4 amd64 4.9.5-0ubuntu0.12.04.2 [141 kB]\nGet:41 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnss3 amd64 3.15.4-0ubuntu0.12.04.2 [1,229 kB]\nGet:42 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libnss3-1d amd64 3.15.4-0ubuntu0.12.04.2 [13.4 kB]\nGet:43 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ca-certificates-java all 20110912ubuntu6 [8,186 B]\nGet:44 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main tzdata-java all 2014a-0ubuntu0.12.04 [126 kB]\nGet:45 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main java-common all 0.43ubuntu2 [61.7 kB]\nGet:46 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liblcms2-2 amd64 2.2+git20110628-2ubuntu3.1 [143 kB]\nGet:47 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jre-headless amd64 7u51-2.4.4-0ubuntu0.12.04.2 [41.4 MB]\nGet:48 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgif4 amd64 4.1.6-9ubuntu1 [31.0 kB]\nGet:49 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libjson0 amd64 0.9-1ubuntu1 [16.3 kB]\nGet:50 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libogg0 amd64 1.2.2~dfsg-1ubuntu1 [17.2 kB]\nGet:51 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libflac8 amd64 1.2.1-6 [144 kB]\nGet:52 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbis0a amd64 1.3.2-1ubuntu3 [103 kB]\nGet:53 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbisenc2 amd64 1.3.2-1ubuntu3 [136 kB]\nGet:54 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsndfile1 amd64 1.0.25-4 [163 kB]\nGet:55 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpulse0 amd64 1:1.1-0ubuntu15.4 [289 kB]\nGet:56 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11-common all 1:7.6+12ubuntu2 [52.0 kB]\nGet:57 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxtst6 amd64 2:1.2.0-4ubuntu0.1 [13.9 kB]\nGet:58 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libglapi-mesa amd64 8.0.4-0ubuntu0.7 [21.3 kB]\nGet:59 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-xcb1 amd64 2:1.4.99.1-0ubuntu2.2 [10.5 kB]\nGet:60 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-glx0 amd64 1.8.1-1ubuntu0.2 [26.4 kB]\nGet:61 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxxf86vm1 amd64 1:1.1.1-2ubuntu0.1 [11.9 kB]\nGet:62 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgl1-mesa-glx amd64 8.0.4-0ubuntu0.7 [104 kB]\nGet:63 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jre amd64 7u51-2.4.4-0ubuntu0.12.04.2 [226 kB]\nGet:64 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk-wrapper-java all 0.30.4-0ubuntu2 [30.9 kB]\nGet:65 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libatk-wrapper-java-jni amd64 0.30.4-0ubuntu2 [30.5 kB]\nGet:66 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libavahi-glib1 amd64 0.6.30-5ubuntu2.1 [8,838 B]\nGet:67 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libbonobo2-common all 2.32.1-0ubuntu1.1 [37.4 kB]\nGet:68 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libmpfr4 amd64 3.1.0-3ubuntu2 [201 kB]\nGet:69 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libmpc2 amd64 0.9-4 [39.5 kB]\nGet:70 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cpp-4.6 amd64 4.6.3-1ubuntu5 [4,836 kB]\nGet:71 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cpp amd64 4:4.6.3-1ubuntu5 [27.6 kB]\nGet:72 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libidl-common all 0.8.14-0.2ubuntu2 [8,572 B]\nGet:73 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libidl0 amd64 0.8.14-0.2ubuntu2 [81.6 kB]\nGet:74 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liborbit2 amd64 1:2.14.19-0.1ubuntu1 [186 kB]\nGet:75 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libbonobo2-0 amd64 2.32.1-0ubuntu1.1 [285 kB]\nGet:76 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libcairo-gobject2 amd64 1.10.2-6.1ubuntu3 [20.5 kB]\nGet:77 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libltdl7 amd64 2.4.2-1ubuntu1 [37.6 kB]\nGet:78 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libtdb1 amd64 1.2.9-4 [35.3 kB]\nGet:79 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libvorbisfile3 amd64 1.3.2-1ubuntu3 [17.7 kB]\nGet:80 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main sound-theme-freedesktop all 0.7.pristine-2 [412 kB]\nGet:81 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libcanberra0 amd64 0.28-3ubuntu3 [43.8 kB]\nGet:82 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf2-common all 3.2.5-0ubuntu2 [22.1 kB]\nGet:83 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgconf-2-4 amd64 3.2.5-0ubuntu2 [97.2 kB]\nGet:84 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgnome-keyring-common all 3.2.2-2 [5,060 B]\nGet:85 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgnome-keyring0 amd64 3.2.2-2 [62.6 kB]\nGet:86 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgudev-1.0-0 amd64 1:175-0ubuntu9.5 [14.5 kB]\nGet:87 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libdevmapper-event1.02.1 amd64 2:1.02.48-4ubuntu7.4 [10.4 kB]\nGet:88 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main liblvm2app2.2 amd64 2.02.66-4ubuntu7.4 [223 kB]\nGet:89 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsgutils2-2 amd64 1.33-1 [64.4 kB]\nGet:90 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main udisks amd64 1.0.4-5ubuntu2.2 [250 kB]\nGet:91 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgdu0 amd64 3.0.2-2ubuntu7 [90.0 kB]\nGet:92 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf-service-backend amd64 3.2.5-0ubuntu2 [74.1 kB]\nGet:93 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf-service amd64 3.2.5-0ubuntu2 [2,012 B]\nGet:94 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dbus-x11 amd64 1.4.18-1ubuntu1.4 [22.1 kB]\nGet:95 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main gconf2 amd64 3.2.5-0ubuntu2 [79.7 kB]\nGet:96 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnomevfs2-common amd64 1:2.24.4-1ubuntu2.1 [24.6 kB]\nGet:97 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnomevfs2-0 amd64 1:2.24.4-1ubuntu2.1 [275 kB]\nGet:98 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-common all 2.32.1-2ubuntu1.1 [36.9 kB]\nGet:99 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-bin amd64 2.32.1-2ubuntu1.1 [15.9 kB]\nGet:100 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgnome2-0 amd64 2.32.1-2ubuntu1.1 [53.4 kB]\nGet:101 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libdconf0 amd64 0.12.0-0ubuntu1.1 [25.4 kB]\nGet:102 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dconf-service amd64 0.12.0-0ubuntu1.1 [16.9 kB]\nGet:103 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main dconf-gsettings-backend amd64 0.12.0-0ubuntu1.1 [16.1 kB]\nGet:104 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-common all 3.4.2-0ubuntu0.8 [145 kB]\nGet:105 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-0 amd64 3.4.2-0ubuntu0.8 [2,281 kB]\nGet:106 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libice6 amd64 2:1.0.7-2build1 [46.1 kB]\nGet:107 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libllvm3.0 amd64 3.0-4ubuntu1 [7,456 kB]\nGet:108 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpolkit-agent-1-0 amd64 0.104-1ubuntu1.1 [17.7 kB]\nGet:109 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpolkit-backend-1-0 amd64 0.104-1ubuntu1.1 [41.7 kB]\nGet:110 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsm6 amd64 2:1.2.0-2build1 [18.1 kB]\nGet:111 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxt6 amd64 1:1.1.1-2ubuntu0.1 [186 kB]\nGet:112 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxmu6 amd64 2:1.1.0-3 [52.7 kB]\nGet:113 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxpm4 amd64 1:3.5.9-4 [37.0 kB]\nGet:114 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxaw7 amd64 2:1.0.9-3ubuntu1 [201 kB]\nGet:115 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb-shape0 amd64 1.8.1-1ubuntu0.2 [6,244 B]\nGet:116 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxv1 amd64 2:1.0.6-2ubuntu0.1 [12.4 kB]\nGet:117 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libck-connector0 amd64 0.4.5-2ubuntu0.1 [10.4 kB]\nGet:118 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main consolekit amd64 0.4.5-2ubuntu0.1 [120 kB]\nGet:119 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libcryptsetup4 amd64 2:1.4.1-2ubuntu4 [55.8 kB]\nGet:120 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main cryptsetup-bin amd64 2:1.4.1-2ubuntu4 [32.2 kB]\nGet:121 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libfontenc1 amd64 1:1.1.0-1 [15.4 kB]\nGet:122 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxxf86dga1 amd64 2:1.1.2-1ubuntu0.1 [13.9 kB]\nGet:123 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11-utils amd64 7.6+4ubuntu0.1 [230 kB]\nGet:124 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-common all 1.12.1-0ubuntu1.2 [16.4 kB]\nGet:125 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-libs amd64 1.12.1-0ubuntu1.2 [52.8 kB]\nGet:126 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs-daemons amd64 1.12.1-0ubuntu1.2 [348 kB]\nGet:127 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main gvfs amd64 1.12.1-0ubuntu1.2 [98.9 kB]\nGet:128 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main hicolor-icon-theme all 0.12-1ubuntu2 [10.2 kB]\nGet:129 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libgconf2-4 amd64 3.2.5-0ubuntu2 [2,016 B]\nGet:130 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgl1-mesa-dri amd64 8.0.4-0ubuntu0.7 [3,035 kB]\nGet:131 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk-3-bin amd64 3.4.2-0ubuntu0.8 [15.9 kB]\nGet:132 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libgtk2.0-bin amd64 2.24.10-0ubuntu6.1 [10.3 kB]\nGet:133 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main xorg-sgml-doctools all 1:1.10-1 [12.0 kB]\nGet:134 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11proto-core-dev all 7.0.22-1ubuntu0.1 [300 kB]\nGet:135 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libice-dev amd64 2:1.0.7-2build1 [133 kB]\nGet:136 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libpam-ck-connector amd64 0.4.5-2ubuntu0.1 [8,188 B]\nGet:137 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libpthread-stubs0 amd64 0.3-3 [3,258 B]\nGet:138 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libpthread-stubs0-dev amd64 0.3-3 [2,866 B]\nGet:139 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libsm-dev amd64 2:1.2.0-2build1 [89.5 kB]\nGet:140 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxau-dev amd64 1:1.0.6-4 [10.5 kB]\nGet:141 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main libxdmcp-dev amd64 1:1.1.0-4 [26.9 kB]\nGet:142 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main x11proto-input-dev all 2.3-1~precise1 [139 kB]\nGet:143 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main x11proto-kb-dev all 1.0.5-2 [27.6 kB]\nGet:144 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main xtrans-dev all 1.2.6-2 [82.9 kB]\nGet:145 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxcb1-dev amd64 1.8.1-1ubuntu0.2 [82.5 kB]\nGet:146 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-dev amd64 2:1.4.99.1-0ubuntu2.2 [912 kB]\nGet:147 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libx11-doc all 2:1.4.99.1-0ubuntu2.2 [2,435 kB]\nGet:148 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main libxt-dev amd64 1:1.1.1-2ubuntu0.1 [490 kB]\nGet:149 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main mtools amd64 4.0.12-1ubuntu0.12.04.1 [186 kB]\nGet:150 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe openjdk-7-jdk amd64 7u51-2.4.4-0ubuntu0.12.04.2 [16.4 MB]\nGet:151 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main policykit-1 amd64 0.104-1ubuntu1.1 [54.9 kB]\nGet:152 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/main policykit-1-gnome amd64 0.105-1ubuntu3.1 [27.7 kB]\nGet:153 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise/main ttf-dejavu-extra all 2.33-2ubuntu1 [3,420 kB]\nGet:154 http://us-east-1.ec2.archive.ubuntu.com/ubuntu/ precise-updates/universe icedtea-7-jre-jamvm amd64 7u51-2.4.4-0ubuntu0.12.04.2 [558 kB]\nFetched 100 MB in 10s (10.0 MB/s)\nExtracting templates from packages: 19%\nExtracting templates from packages: 38%\nExtracting templates from packages: 58%\nExtracting templates from packages: 77%\nExtracting templates from packages: 97%\nExtracting templates from packages: 100%\nPreconfiguring packages ...\nSelecting previously unselected package libasound2.\n(Reading database ... \n(Reading database ... 5%\n(Reading database ... 10%\n(Reading database ... 15%\n(Reading database ... 20%\n(Reading database ... 25%\n(Reading database ... 30%\n(Reading database ... 35%\n(Reading database ... 40%\n(Reading database ... 45%\n(Reading database ... 50%\n(Reading database ... 55%\n(Reading database ... 60%\n(Reading database ... 65%\n(Reading database ... 70%\n(Reading database ... 75%\n(Reading database ... 80%\n(Reading database ... 85%\n(Reading database ... 90%\n(Reading database ... 95%\n(Reading database ... 100%\n(Reading database ... 47578 files and directories currently installed.)\nUnpacking libasound2 (from .../libasound2_1.0.25-1ubuntu10.2_amd64.deb) ...\nSelecting previously unselected package libasyncns0.\nUnpacking libasyncns0 (from .../libasyncns0_0.8-4_amd64.deb) ...\nSelecting previously unselected package libatasmart4.\nUnpacking libatasmart4 (from .../libatasmart4_0.18-3_amd64.deb) ...\nSelecting previously unselected package libatk1.0-data.\nUnpacking libatk1.0-data (from .../libatk1.0-data_2.4.0-0ubuntu1_all.deb) ...\nSelecting previously unselected package libatk1.0-0.\nUnpacking libatk1.0-0 (from .../libatk1.0-0_2.4.0-0ubuntu1_amd64.deb) ...\nSelecting previously unselected package libgtk2.0-common.\nUnpacking libgtk2.0-common (from .../libgtk2.0-common_2.24.10-0ubuntu6.1_all.deb) ...\nSelecting previously unselected package ttf-dejavu-core.\nUnpacking ttf-dejavu-core (from .../ttf-dejavu-core_2.33-2ubuntu1_all.deb) ...\nSelecting previously unselected package fontconfig-config.\nUnpacking fontconfig-config (from .../fontconfig-config_2.8.0-3ubuntu9.1_all.deb) ...\nSelecting previously unselected package libfontconfig1.\nUnpacking libfontconfig1 (from .../libfontconfig1_2.8.0-3ubuntu9.1_amd64.deb) ...\nSelecting previously unselected package libpixman-1-0.\nUnpacking libpixman-1-0 (from .../libpixman-1-0_0.30.2-1ubuntu0.0.0.0.1_amd64.deb) ...\nSelecting previously unselected package libxcb-render0.\nUnpacking libxcb-render0 (from .../libxcb-render0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxcb-shm0.\nUnpacking libxcb-shm0 (from .../libxcb-shm0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxrender1.\nUnpacking libxrender1 (from .../libxrender1_1%3a0.9.6-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libcairo2.\nUnpacking libcairo2 (from .../libcairo2_1.10.2-6.1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libavahi-common-data.\nUnpacking libavahi-common-data (from .../libavahi-common-data_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libavahi-common3.\nUnpacking libavahi-common3 (from .../libavahi-common3_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libavahi-client3.\nUnpacking libavahi-client3 (from .../libavahi-client3_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libcups2.\nUnpacking libcups2 (from .../libcups2_1.5.3-0ubuntu8.2_amd64.deb) ...\nSelecting previously unselected package libjpeg-turbo8.\nUnpacking libjpeg-turbo8 (from .../libjpeg-turbo8_1.1.90+svn733-0ubuntu4.3_amd64.deb) ...\nSelecting previously unselected package libjpeg8.\nUnpacking libjpeg8 (from .../libjpeg8_8c-2ubuntu7_amd64.deb) ...\nSelecting previously unselected package libjasper1.\nUnpacking libjasper1 (from .../libjasper1_1.900.1-13_amd64.deb) ...\nSelecting previously unselected package libtiff4.\nUnpacking libtiff4 (from .../libtiff4_3.9.5-2ubuntu1.5_amd64.deb) ...\nSelecting previously unselected package libgdk-pixbuf2.0-common.\nUnpacking libgdk-pixbuf2.0-common (from .../libgdk-pixbuf2.0-common_2.26.1-1ubuntu1.1_all.deb) ...\nSelecting previously unselected package libgdk-pixbuf2.0-0.\nUnpacking libgdk-pixbuf2.0-0 (from .../libgdk-pixbuf2.0-0_2.26.1-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libthai-data.\nUnpacking libthai-data (from .../libthai-data_0.1.16-3_all.deb) ...\nSelecting previously unselected package libdatrie1.\nUnpacking libdatrie1 (from .../libdatrie1_0.2.5-3_amd64.deb) ...\nSelecting previously unselected package libthai0.\nUnpacking libthai0 (from .../libthai0_0.1.16-3_amd64.deb) ...\nSelecting previously unselected package libxft2.\nUnpacking libxft2 (from .../libxft2_2.2.0-3ubuntu2_amd64.deb) ...\nSelecting previously unselected package fontconfig.\nUnpacking fontconfig (from .../fontconfig_2.8.0-3ubuntu9.1_amd64.deb) ...\nSelecting previously unselected package libpango1.0-0.\nUnpacking libpango1.0-0 (from .../libpango1.0-0_1.30.0-0ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package libxcomposite1.\nUnpacking libxcomposite1 (from .../libxcomposite1_1%3a0.4.3-2build1_amd64.deb) ...\nSelecting previously unselected package libxfixes3.\nUnpacking libxfixes3 (from .../libxfixes3_1%3a5.0-4ubuntu4.2_amd64.deb) ...\nSelecting previously unselected package libxcursor1.\nUnpacking libxcursor1 (from .../libxcursor1_1%3a1.1.12-1ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxdamage1.\nUnpacking libxdamage1 (from .../libxdamage1_1%3a1.1.3-2build1_amd64.deb) ...\nSelecting previously unselected package libxi6.\nUnpacking libxi6 (from .../libxi6_2%3a1.7.1.901-1ubuntu1~precise1_amd64.deb) ...\nSelecting previously unselected package libxinerama1.\nUnpacking libxinerama1 (from .../libxinerama1_2%3a1.1.1-3ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxrandr2.\nUnpacking libxrandr2 (from .../libxrandr2_2%3a1.3.2-2ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package shared-mime-info.\nUnpacking shared-mime-info (from .../shared-mime-info_1.0-0ubuntu4.1_amd64.deb) ...\nSelecting previously unselected package libgtk2.0-0.\nUnpacking libgtk2.0-0 (from .../libgtk2.0-0_2.24.10-0ubuntu6.1_amd64.deb) ...\nSelecting previously unselected package libnspr4.\nUnpacking libnspr4 (from .../libnspr4_4.9.5-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libnss3.\nUnpacking libnss3 (from .../libnss3_3.15.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libnss3-1d.\nUnpacking libnss3-1d (from .../libnss3-1d_3.15.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package ca-certificates-java.\nUnpacking ca-certificates-java (from .../ca-certificates-java_20110912ubuntu6_all.deb) ...\nSelecting previously unselected package tzdata-java.\nUnpacking tzdata-java (from .../tzdata-java_2014a-0ubuntu0.12.04_all.deb) ...\nSelecting previously unselected package java-common.\nUnpacking java-common (from .../java-common_0.43ubuntu2_all.deb) ...\nSelecting previously unselected package liblcms2-2.\nUnpacking liblcms2-2 (from .../liblcms2-2_2.2+git20110628-2ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jre-headless.\nUnpacking openjdk-7-jre-headless (from .../openjdk-7-jre-headless_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libgif4.\nUnpacking libgif4 (from .../libgif4_4.1.6-9ubuntu1_amd64.deb) ...\nSelecting previously unselected package libjson0.\nUnpacking libjson0 (from .../libjson0_0.9-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libogg0.\nUnpacking libogg0 (from .../libogg0_1.2.2~dfsg-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libflac8.\nUnpacking libflac8 (from .../libflac8_1.2.1-6_amd64.deb) ...\nSelecting previously unselected package libvorbis0a.\nUnpacking libvorbis0a (from .../libvorbis0a_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libvorbisenc2.\nUnpacking libvorbisenc2 (from .../libvorbisenc2_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libsndfile1.\nUnpacking libsndfile1 (from .../libsndfile1_1.0.25-4_amd64.deb) ...\nSelecting previously unselected package libpulse0.\nUnpacking libpulse0 (from .../libpulse0_1%3a1.1-0ubuntu15.4_amd64.deb) ...\nSelecting previously unselected package x11-common.\nUnpacking x11-common (from .../x11-common_1%3a7.6+12ubuntu2_all.deb) ...\nSelecting previously unselected package libxtst6.\nUnpacking libxtst6 (from .../libxtst6_2%3a1.2.0-4ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libglapi-mesa.\nUnpacking libglapi-mesa (from .../libglapi-mesa_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package libx11-xcb1.\nUnpacking libx11-xcb1 (from .../libx11-xcb1_2%3a1.4.99.1-0ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libxcb-glx0.\nUnpacking libxcb-glx0 (from .../libxcb-glx0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxxf86vm1.\nUnpacking libxxf86vm1 (from .../libxxf86vm1_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libgl1-mesa-glx.\nUnpacking libgl1-mesa-glx (from .../libgl1-mesa-glx_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jre.\nUnpacking openjdk-7-jre (from .../openjdk-7-jre_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package libatk-wrapper-java.\nUnpacking libatk-wrapper-java (from .../libatk-wrapper-java_0.30.4-0ubuntu2_all.deb) ...\nSelecting previously unselected package libatk-wrapper-java-jni.\nUnpacking libatk-wrapper-java-jni (from .../libatk-wrapper-java-jni_0.30.4-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libavahi-glib1.\nUnpacking libavahi-glib1 (from .../libavahi-glib1_0.6.30-5ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libbonobo2-common.\nUnpacking libbonobo2-common (from .../libbonobo2-common_2.32.1-0ubuntu1.1_all.deb) ...\nSelecting previously unselected package libmpfr4.\nUnpacking libmpfr4 (from .../libmpfr4_3.1.0-3ubuntu2_amd64.deb) ...\nSelecting previously unselected package libmpc2.\nUnpacking libmpc2 (from .../libmpc2_0.9-4_amd64.deb) ...\nSelecting previously unselected package cpp-4.6.\nUnpacking cpp-4.6 (from .../cpp-4.6_4.6.3-1ubuntu5_amd64.deb) ...\nSelecting previously unselected package cpp.\nUnpacking cpp (from .../cpp_4%3a4.6.3-1ubuntu5_amd64.deb) ...\nSelecting previously unselected package libidl-common.\nUnpacking libidl-common (from .../libidl-common_0.8.14-0.2ubuntu2_all.deb) ...\nSelecting previously unselected package libidl0.\nUnpacking libidl0 (from .../libidl0_0.8.14-0.2ubuntu2_amd64.deb) ...\nSelecting previously unselected package liborbit2.\nUnpacking liborbit2 (from .../liborbit2_1%3a2.14.19-0.1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libbonobo2-0.\nUnpacking libbonobo2-0 (from .../libbonobo2-0_2.32.1-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libcairo-gobject2.\nUnpacking libcairo-gobject2 (from .../libcairo-gobject2_1.10.2-6.1ubuntu3_amd64.deb) ...\nSelecting previously unselected package libltdl7.\nUnpacking libltdl7 (from .../libltdl7_2.4.2-1ubuntu1_amd64.deb) ...\nSelecting previously unselected package libtdb1.\nUnpacking libtdb1 (from .../libtdb1_1.2.9-4_amd64.deb) ...\nSelecting previously unselected package libvorbisfile3.\nUnpacking libvorbisfile3 (from .../libvorbisfile3_1.3.2-1ubuntu3_amd64.deb) ...\nSelecting previously unselected package sound-theme-freedesktop.\nUnpacking sound-theme-freedesktop (from .../sound-theme-freedesktop_0.7.pristine-2_all.deb) ...\nSelecting previously unselected package libcanberra0.\nUnpacking libcanberra0 (from .../libcanberra0_0.28-3ubuntu3_amd64.deb) ...\nSelecting previously unselected package gconf2-common.\nUnpacking gconf2-common (from .../gconf2-common_3.2.5-0ubuntu2_all.deb) ...\nSelecting previously unselected package libgconf-2-4.\nUnpacking libgconf-2-4 (from .../libgconf-2-4_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgnome-keyring-common.\nUnpacking libgnome-keyring-common (from .../libgnome-keyring-common_3.2.2-2_all.deb) ...\nSelecting previously unselected package libgnome-keyring0.\nUnpacking libgnome-keyring0 (from .../libgnome-keyring0_3.2.2-2_amd64.deb) ...\nSelecting previously unselected package libgudev-1.0-0.\nUnpacking libgudev-1.0-0 (from .../libgudev-1.0-0_1%3a175-0ubuntu9.5_amd64.deb) ...\nSelecting previously unselected package libdevmapper-event1.02.1.\nUnpacking libdevmapper-event1.02.1 (from .../libdevmapper-event1.02.1_2%3a1.02.48-4ubuntu7.4_amd64.deb) ...\nSelecting previously unselected package liblvm2app2.2.\nUnpacking liblvm2app2.2 (from .../liblvm2app2.2_2.02.66-4ubuntu7.4_amd64.deb) ...\nSelecting previously unselected package libsgutils2-2.\nUnpacking libsgutils2-2 (from .../libsgutils2-2_1.33-1_amd64.deb) ...\nSelecting previously unselected package udisks.\nUnpacking udisks (from .../udisks_1.0.4-5ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libgdu0.\nUnpacking libgdu0 (from .../libgdu0_3.0.2-2ubuntu7_amd64.deb) ...\nSelecting previously unselected package gconf-service-backend.\nUnpacking gconf-service-backend (from .../gconf-service-backend_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package gconf-service.\nUnpacking gconf-service (from .../gconf-service_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package dbus-x11.\nUnpacking dbus-x11 (from .../dbus-x11_1.4.18-1ubuntu1.4_amd64.deb) ...\nSelecting previously unselected package gconf2.\nUnpacking gconf2 (from .../gconf2_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgnomevfs2-common.\nUnpacking libgnomevfs2-common (from .../libgnomevfs2-common_1%3a2.24.4-1ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libgnomevfs2-0.\nUnpacking libgnomevfs2-0 (from .../libgnomevfs2-0_1%3a2.24.4-1ubuntu2.1_amd64.deb) ...\nSelecting previously unselected package libgnome2-common.\nUnpacking libgnome2-common (from .../libgnome2-common_2.32.1-2ubuntu1.1_all.deb) ...\nSelecting previously unselected package libgnome2-bin.\nUnpacking libgnome2-bin (from .../libgnome2-bin_2.32.1-2ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libgnome2-0.\nUnpacking libgnome2-0 (from .../libgnome2-0_2.32.1-2ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libdconf0.\nUnpacking libdconf0 (from .../libdconf0_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package dconf-service.\nUnpacking dconf-service (from .../dconf-service_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package dconf-gsettings-backend.\nUnpacking dconf-gsettings-backend (from .../dconf-gsettings-backend_0.12.0-0ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libgtk-3-common.\nUnpacking libgtk-3-common (from .../libgtk-3-common_3.4.2-0ubuntu0.8_all.deb) ...\nSelecting previously unselected package libgtk-3-0.\nUnpacking libgtk-3-0 (from .../libgtk-3-0_3.4.2-0ubuntu0.8_amd64.deb) ...\nSelecting previously unselected package libice6.\nUnpacking libice6 (from .../libice6_2%3a1.0.7-2build1_amd64.deb) ...\nSelecting previously unselected package libllvm3.0.\nUnpacking libllvm3.0 (from .../libllvm3.0_3.0-4ubuntu1_amd64.deb) ...\nSelecting previously unselected package libpolkit-agent-1-0.\nUnpacking libpolkit-agent-1-0 (from .../libpolkit-agent-1-0_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libpolkit-backend-1-0.\nUnpacking libpolkit-backend-1-0 (from .../libpolkit-backend-1-0_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package libsm6.\nUnpacking libsm6 (from .../libsm6_2%3a1.2.0-2build1_amd64.deb) ...\nSelecting previously unselected package libxt6.\nUnpacking libxt6 (from .../libxt6_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libxmu6.\nUnpacking libxmu6 (from .../libxmu6_2%3a1.1.0-3_amd64.deb) ...\nSelecting previously unselected package libxpm4.\nUnpacking libxpm4 (from .../libxpm4_1%3a3.5.9-4_amd64.deb) ...\nSelecting previously unselected package libxaw7.\nUnpacking libxaw7 (from .../libxaw7_2%3a1.0.9-3ubuntu1_amd64.deb) ...\nSelecting previously unselected package libxcb-shape0.\nUnpacking libxcb-shape0 (from .../libxcb-shape0_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libxv1.\nUnpacking libxv1 (from .../libxv1_2%3a1.0.6-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libck-connector0.\nUnpacking libck-connector0 (from .../libck-connector0_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package consolekit.\nUnpacking consolekit (from .../consolekit_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libcryptsetup4.\nUnpacking libcryptsetup4 (from .../libcryptsetup4_2%3a1.4.1-2ubuntu4_amd64.deb) ...\nSelecting previously unselected package cryptsetup-bin.\nUnpacking cryptsetup-bin (from .../cryptsetup-bin_2%3a1.4.1-2ubuntu4_amd64.deb) ...\nSelecting previously unselected package libfontenc1.\nUnpacking libfontenc1 (from .../libfontenc1_1%3a1.1.0-1_amd64.deb) ...\nSelecting previously unselected package libxxf86dga1.\nUnpacking libxxf86dga1 (from .../libxxf86dga1_2%3a1.1.2-1ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package x11-utils.\nUnpacking x11-utils (from .../x11-utils_7.6+4ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package gvfs-common.\nUnpacking gvfs-common (from .../gvfs-common_1.12.1-0ubuntu1.2_all.deb) ...\nSelecting previously unselected package gvfs-libs.\nUnpacking gvfs-libs (from .../gvfs-libs_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package gvfs-daemons.\nUnpacking gvfs-daemons (from .../gvfs-daemons_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package gvfs.\nUnpacking gvfs (from .../gvfs_1.12.1-0ubuntu1.2_amd64.deb) ...\nSelecting previously unselected package hicolor-icon-theme.\nUnpacking hicolor-icon-theme (from .../hicolor-icon-theme_0.12-1ubuntu2_all.deb) ...\nSelecting previously unselected package libgconf2-4.\nUnpacking libgconf2-4 (from .../libgconf2-4_3.2.5-0ubuntu2_amd64.deb) ...\nSelecting previously unselected package libgl1-mesa-dri.\nUnpacking libgl1-mesa-dri (from .../libgl1-mesa-dri_8.0.4-0ubuntu0.7_amd64.deb) ...\nSelecting previously unselected package libgtk-3-bin.\nUnpacking libgtk-3-bin (from .../libgtk-3-bin_3.4.2-0ubuntu0.8_amd64.deb) ...\nAdding 'diversion of /usr/sbin/update-icon-caches to /usr/sbin/update-icon-caches.gtk2 by libgtk-3-bin'\nAdding 'diversion of /usr/share/man/man8/update-icon-caches.8.gz to /usr/share/man/man8/update-icon-caches.gtk2.8.gz by libgtk-3-bin'\nSelecting previously unselected package libgtk2.0-bin.\nUnpacking libgtk2.0-bin (from .../libgtk2.0-bin_2.24.10-0ubuntu6.1_amd64.deb) ...\nSelecting previously unselected package xorg-sgml-doctools.\nUnpacking xorg-sgml-doctools (from .../xorg-sgml-doctools_1%3a1.10-1_all.deb) ...\nSelecting previously unselected package x11proto-core-dev.\nUnpacking x11proto-core-dev (from .../x11proto-core-dev_7.0.22-1ubuntu0.1_all.deb) ...\nSelecting previously unselected package libice-dev.\nUnpacking libice-dev (from .../libice-dev_2%3a1.0.7-2build1_amd64.deb) ...\nSelecting previously unselected package libpam-ck-connector.\nUnpacking libpam-ck-connector (from .../libpam-ck-connector_0.4.5-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package libpthread-stubs0.\nUnpacking libpthread-stubs0 (from .../libpthread-stubs0_0.3-3_amd64.deb) ...\nSelecting previously unselected package libpthread-stubs0-dev.\nUnpacking libpthread-stubs0-dev (from .../libpthread-stubs0-dev_0.3-3_amd64.deb) ...\nSelecting previously unselected package libsm-dev.\nUnpacking libsm-dev (from .../libsm-dev_2%3a1.2.0-2build1_amd64.deb) ...\nSelecting previously unselected package libxau-dev.\nUnpacking libxau-dev (from .../libxau-dev_1%3a1.0.6-4_amd64.deb) ...\nSelecting previously unselected package libxdmcp-dev.\nUnpacking libxdmcp-dev (from .../libxdmcp-dev_1%3a1.1.0-4_amd64.deb) ...\nSelecting previously unselected package x11proto-input-dev.\nUnpacking x11proto-input-dev (from .../x11proto-input-dev_2.3-1~precise1_all.deb) ...\nSelecting previously unselected package x11proto-kb-dev.\nUnpacking x11proto-kb-dev (from .../x11proto-kb-dev_1.0.5-2_all.deb) ...\nSelecting previously unselected package xtrans-dev.\nUnpacking xtrans-dev (from .../xtrans-dev_1.2.6-2_all.deb) ...\nSelecting previously unselected package libxcb1-dev.\nUnpacking libxcb1-dev (from .../libxcb1-dev_1.8.1-1ubuntu0.2_amd64.deb) ...\nSelecting previously unselected package libx11-dev.\nUnpacking libx11-dev (from .../libx11-dev_2%3a1.4.99.1-0ubuntu2.2_amd64.deb) ...\nSelecting previously unselected package libx11-doc.\nUnpacking libx11-doc (from .../libx11-doc_2%3a1.4.99.1-0ubuntu2.2_all.deb) ...\nSelecting previously unselected package libxt-dev.\nUnpacking libxt-dev (from .../libxt-dev_1%3a1.1.1-2ubuntu0.1_amd64.deb) ...\nSelecting previously unselected package mtools.\nUnpacking mtools (from .../mtools_4.0.12-1ubuntu0.12.04.1_amd64.deb) ...\nSelecting previously unselected package openjdk-7-jdk.\nUnpacking openjdk-7-jdk (from .../openjdk-7-jdk_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nSelecting previously unselected package policykit-1.\nUnpacking policykit-1 (from .../policykit-1_0.104-1ubuntu1.1_amd64.deb) ...\nSelecting previously unselected package policykit-1-gnome.\nUnpacking policykit-1-gnome (from .../policykit-1-gnome_0.105-1ubuntu3.1_amd64.deb) ...\nSelecting previously unselected package ttf-dejavu-extra.\nUnpacking ttf-dejavu-extra (from .../ttf-dejavu-extra_2.33-2ubuntu1_all.deb) ...\nSelecting previously unselected package icedtea-7-jre-jamvm.\nUnpacking icedtea-7-jre-jamvm (from .../icedtea-7-jre-jamvm_7u51-2.4.4-0ubuntu0.12.04.2_amd64.deb) ...\nProcessing triggers for initramfs-tools ...\nupdate-initramfs: Generating /boot/initrd.img-3.2.0-60-virtual\nProcessing triggers for man-db ...\nProcessing triggers for ureadahead ...\nProcessing triggers for libglib2.0-0 ...\nProcessing triggers for install-info ...\nSetting up libasound2 (1.0.25-1ubuntu10.2) ...\nSetting up libasyncns0 (0.8-4) ...\nSetting up libatasmart4 (0.18-3) ...\nSetting up libatk1.0-data (2.4.0-0ubuntu1) ...\nSetting up libatk1.0-0 (2.4.0-0ubuntu1) ...\nSetting up libgtk2.0-common (2.24.10-0ubuntu6.1) ...\nSetting up ttf-dejavu-core (2.33-2ubuntu1) ...\nSetting up fontconfig-config (2.8.0-3ubuntu9.1) ...\nSetting up libfontconfig1 (2.8.0-3ubuntu9.1) ...\nSetting up libpixman-1-0 (0.30.2-1ubuntu0.0.0.0.1) ...\nSetting up libxcb-render0 (1.8.1-1ubuntu0.2) ...\nSetting up libxcb-shm0 (1.8.1-1ubuntu0.2) ...\nSetting up libxrender1 (1:0.9.6-2ubuntu0.1) ...\nSetting up libcairo2 (1.10.2-6.1ubuntu3) ...\nSetting up libavahi-common-data (0.6.30-5ubuntu2.1) ...\nSetting up libavahi-common3 (0.6.30-5ubuntu2.1) ...\nSetting up libavahi-client3 (0.6.30-5ubuntu2.1) ...\nSetting up libcups2 (1.5.3-0ubuntu8.2) ...\nSetting up libjpeg-turbo8 (1.1.90+svn733-0ubuntu4.3) ...\nSetting up libjpeg8 (8c-2ubuntu7) ...\nSetting up libjasper1 (1.900.1-13) ...\nSetting up libtiff4 (3.9.5-2ubuntu1.5) ...\nSetting up libgdk-pixbuf2.0-common (2.26.1-1ubuntu1.1) ...\nSetting up libgdk-pixbuf2.0-0 (2.26.1-1ubuntu1.1) ...\nSetting up libthai-data (0.1.16-3) ...\nSetting up libdatrie1 (0.2.5-3) ...\nSetting up libthai0 (0.1.16-3) ...\nSetting up libxft2 (2.2.0-3ubuntu2) ...\nSetting up fontconfig (2.8.0-3ubuntu9.1) ...\nCleaning up old fontconfig caches... done.\nRegenerating fonts cache... done.\nSetting up libpango1.0-0 (1.30.0-0ubuntu3.1) ...\nSetting up libxcomposite1 (1:0.4.3-2build1) ...\nSetting up libxfixes3 (1:5.0-4ubuntu4.2) ...\nSetting up libxcursor1 (1:1.1.12-1ubuntu0.1) ...\nSetting up libxdamage1 (1:1.1.3-2build1) ...\nSetting up libxi6 (2:1.7.1.901-1ubuntu1~precise1) ...\nSetting up libxinerama1 (2:1.1.1-3ubuntu0.1) ...\nSetting up libxrandr2 (2:1.3.2-2ubuntu0.2) ...\nSetting up shared-mime-info (1.0-0ubuntu4.1) ...\nSetting up libgtk2.0-0 (2.24.10-0ubuntu6.1) ...\nSetting up libnspr4 (4.9.5-0ubuntu0.12.04.2) ...\nSetting up libnss3 (3.15.4-0ubuntu0.12.04.2) ...\nSetting up libnss3-1d (3.15.4-0ubuntu0.12.04.2) ...\nSetting up tzdata-java (2014a-0ubuntu0.12.04) ...\nSetting up java-common (0.43ubuntu2) ...\nSetting up liblcms2-2 (2.2+git20110628-2ubuntu3.1) ...\nSetting up libgif4 (4.1.6-9ubuntu1) ...\nSetting up libjson0 (0.9-1ubuntu1) ...\nSetting up libogg0 (1.2.2~dfsg-1ubuntu1) ...\nSetting up libflac8 (1.2.1-6) ...\nSetting up libvorbis0a (1.3.2-1ubuntu3) ...\nSetting up libvorbisenc2 (1.3.2-1ubuntu3) ...\nSetting up libsndfile1 (1.0.25-4) ...\nSetting up libpulse0 (1:1.1-0ubuntu15.4) ...\nSetting up x11-common (1:7.6+12ubuntu2) ...\nSetting up libxtst6 (2:1.2.0-4ubuntu0.1) ...\nSetting up libglapi-mesa (8.0.4-0ubuntu0.7) ...\nSetting up libx11-xcb1 (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libxcb-glx0 (1.8.1-1ubuntu0.2) ...\nSetting up libxxf86vm1 (1:1.1.1-2ubuntu0.1) ...\nSetting up libgl1-mesa-glx (8.0.4-0ubuntu0.7) ...\nupdate-alternatives: using /usr/lib/x86_64-linux-gnu/mesa/ld.so.conf to provide /etc/ld.so.conf.d/x86_64-linux-gnu_GL.conf (x86_64-linux-gnu_gl_conf) in auto mode.\nSetting up libavahi-glib1 (0.6.30-5ubuntu2.1) ...\nSetting up libbonobo2-common (2.32.1-0ubuntu1.1) ...\nSetting up libmpfr4 (3.1.0-3ubuntu2) ...\nSetting up libmpc2 (0.9-4) ...\nSetting up cpp-4.6 (4.6.3-1ubuntu5) ...\nSetting up cpp (4:4.6.3-1ubuntu5) ...\nSetting up libidl-common (0.8.14-0.2ubuntu2) ...\nSetting up libidl0 (0.8.14-0.2ubuntu2) ...\nSetting up liborbit2 (1:2.14.19-0.1ubuntu1) ...\nSetting up libbonobo2-0 (2.32.1-0ubuntu1.1) ...\nSetting up libcairo-gobject2 (1.10.2-6.1ubuntu3) ...\nSetting up libltdl7 (2.4.2-1ubuntu1) ...\nSetting up libtdb1 (1.2.9-4) ...\nSetting up libvorbisfile3 (1.3.2-1ubuntu3) ...\nSetting up sound-theme-freedesktop (0.7.pristine-2) ...\nSetting up libcanberra0 (0.28-3ubuntu3) ...\nSetting up gconf2-common (3.2.5-0ubuntu2) ...\nCreating config file /etc/gconf/2/path with new version\nSetting up libgconf-2-4 (3.2.5-0ubuntu2) ...\nSetting up libgnome-keyring-common (3.2.2-2) ...\nSetting up libgnome-keyring0 (3.2.2-2) ...\nSetting up libgudev-1.0-0 (1:175-0ubuntu9.5) ...\nSetting up libdevmapper-event1.02.1 (2:1.02.48-4ubuntu7.4) ...\nSetting up liblvm2app2.2 (2.02.66-4ubuntu7.4) ...\nSetting up libsgutils2-2 (1.33-1) ...\nSetting up udisks (1.0.4-5ubuntu2.2) ...\nSetting up libgdu0 (3.0.2-2ubuntu7) ...\nSetting up dbus-x11 (1.4.18-1ubuntu1.4) ...\nSetting up libdconf0 (0.12.0-0ubuntu1.1) ...\nSetting up dconf-service (0.12.0-0ubuntu1.1) ...\nSetting up dconf-gsettings-backend (0.12.0-0ubuntu1.1) ...\nSetting up libgtk-3-common (3.4.2-0ubuntu0.8) ...\nSetting up libgtk-3-0 (3.4.2-0ubuntu0.8) ...\nSetting up libice6 (2:1.0.7-2build1) ...\nSetting up libllvm3.0 (3.0-4ubuntu1) ...\nSetting up libpolkit-agent-1-0 (0.104-1ubuntu1.1) ...\nSetting up libpolkit-backend-1-0 (0.104-1ubuntu1.1) ...\nSetting up libsm6 (2:1.2.0-2build1) ...\nSetting up libxt6 (1:1.1.1-2ubuntu0.1) ...\nSetting up libxmu6 (2:1.1.0-3) ...\nSetting up libxpm4 (1:3.5.9-4) ...\nSetting up libxaw7 (2:1.0.9-3ubuntu1) ...\nSetting up libxcb-shape0 (1.8.1-1ubuntu0.2) ...\nSetting up libxv1 (2:1.0.6-2ubuntu0.1) ...\nSetting up libck-connector0 (0.4.5-2ubuntu0.1) ...\nSetting up consolekit (0.4.5-2ubuntu0.1) ...\nSetting up libcryptsetup4 (2:1.4.1-2ubuntu4) ...\nSetting up cryptsetup-bin (2:1.4.1-2ubuntu4) ...\nSetting up libfontenc1 (1:1.1.0-1) ...\nSetting up libxxf86dga1 (2:1.1.2-1ubuntu0.1) ...\nSetting up x11-utils (7.6+4ubuntu0.1) ...\nSetting up gvfs-common (1.12.1-0ubuntu1.2) ...\nSetting up gvfs-libs (1.12.1-0ubuntu1.2) ...\nSetting up gvfs-daemons (1.12.1-0ubuntu1.2) ...\nSetting up gvfs (1.12.1-0ubuntu1.2) ...\nSetting up hicolor-icon-theme (0.12-1ubuntu2) ...\nSetting up libgl1-mesa-dri (8.0.4-0ubuntu0.7) ...\nSetting up libgtk-3-bin (3.4.2-0ubuntu0.8) ...\nSetting up libgtk2.0-bin (2.24.10-0ubuntu6.1) ...\nSetting up xorg-sgml-doctools (1:1.10-1) ...\nSetting up x11proto-core-dev (7.0.22-1ubuntu0.1) ...\nSetting up libice-dev (2:1.0.7-2build1) ...\nSetting up libpam-ck-connector (0.4.5-2ubuntu0.1) ...\nSetting up libpthread-stubs0 (0.3-3) ...\nSetting up libpthread-stubs0-dev (0.3-3) ...\nSetting up libsm-dev (2:1.2.0-2build1) ...\nSetting up libxau-dev (1:1.0.6-4) ...\nSetting up libxdmcp-dev (1:1.1.0-4) ...\nSetting up x11proto-input-dev (2.3-1~precise1) ...\nSetting up x11proto-kb-dev (1.0.5-2) ...\nSetting up xtrans-dev (1.2.6-2) ...\nSetting up libxcb1-dev (1.8.1-1ubuntu0.2) ...\nSetting up libx11-dev (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libx11-doc (2:1.4.99.1-0ubuntu2.2) ...\nSetting up libxt-dev (1:1.1.1-2ubuntu0.1) ...\nSetting up mtools (4.0.12-1ubuntu0.12.04.1) ...\nSetting up policykit-1 (0.104-1ubuntu1.1) ...\nSetting up policykit-1-gnome (0.105-1ubuntu3.1) ...\nSetting up ttf-dejavu-extra (2.33-2ubuntu1) ...\nSetting up gconf-service (3.2.5-0ubuntu2) ...\nSetting up libgconf2-4 (3.2.5-0ubuntu2) ...\nSetting up openjdk-7-jre-headless (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/java to provide /usr/bin/java (java) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/keytool to provide /usr/bin/keytool (keytool) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/pack200 to provide /usr/bin/pack200 (pack200) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/rmid to provide /usr/bin/rmid (rmid) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/rmiregistry to provide /usr/bin/rmiregistry (rmiregistry) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/unpack200 to provide /usr/bin/unpack200 (unpack200) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/orbd to provide /usr/bin/orbd (orbd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/servertool to provide /usr/bin/servertool (servertool) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/tnameserv to provide /usr/bin/tnameserv (tnameserv) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/lib/jexec to provide /usr/bin/jexec (jexec) in auto mode.\nSetting up libatk-wrapper-java (0.30.4-0ubuntu2) ...\nSetting up gconf-service-backend (3.2.5-0ubuntu2) ...\nSetting up gconf2 (3.2.5-0ubuntu2) ...\nupdate-alternatives: using /usr/bin/gconftool-2 to provide /usr/bin/gconftool (gconftool) in auto mode.\nSetting up libgnomevfs2-common (1:2.24.4-1ubuntu2.1) ...\nSetting up libgnomevfs2-0 (1:2.24.4-1ubuntu2.1) ...\nSetting up libgnome2-common (2.32.1-2ubuntu1.1) ...\nSetting up libgnome2-bin (2.32.1-2ubuntu1.1) ...\nSetting up libgnome2-0 (2.32.1-2ubuntu1.1) ...\nSetting up icedtea-7-jre-jamvm (7u51-2.4.4-0ubuntu0.12.04.2) ...\nSetting up ca-certificates-java (20110912ubuntu6) ...\nAdding debian:Digital_Signature_Trust_Co._Global_CA_1.pem\nAdding debian:Wells_Fargo_Root_CA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority.pem\nAdding debian:Staat_der_Nederlanden_Root_CA.pem\nAdding debian:ACEDICOM_Root.pem\nAdding debian:Secure_Global_CA.pem\nAdding debian:Entrust.net_Premium_2048_Secure_Server_CA.pem\nAdding debian:NetLock_Business_=Class_B=_Root.pem\nAdding debian:Comodo_Secure_Services_root.pem\nAdding debian:AffirmTrust_Premium.pem\nAdding debian:StartCom_Certification_Authority_G2.pem\nAdding debian:GlobalSign_Root_CA_-_R2.pem\nAdding debian:TC_TrustCenter_Class_2_CA_II.pem\nAdding debian:China_Internet_Network_Information_Center_EV_Certificates_Root.pem\nAdding debian:SwissSign_Gold_CA_-_G2.pem\nAdding debian:NetLock_Notary_=Class_A=_Root.pem\nAdding debian:AddTrust_Public_Services_Root.pem\nAdding debian:GeoTrust_Primary_Certification_Authority.pem\nAdding debian:Starfield_Root_Certificate_Authority_-_G2.pem\nAdding debian:DigiCert_Global_Root_CA.pem\nAdding debian:Firmaprofesional_Root_CA.pem\nAdding debian:D-TRUST_Root_Class_3_CA_2_2009.pem\nAdding debian:GeoTrust_Universal_CA_2.pem\nAdding debian:DigiCert_High_Assurance_EV_Root_CA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:VeriSign_Class_3_Public_Primary_Certification_Authority_-_G4.pem\nAdding debian:Izenpe.com.pem\nAdding debian:Verisign_Class_2_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:A-Trust-nQual-03.pem\nAdding debian:Swisscom_Root_CA_1.pem\nAdding debian:Buypass_Class_2_CA_1.pem\nAdding debian:CA_Disig_Root_R2.pem\nAdding debian:QuoVadis_Root_CA.pem\nAdding debian:Network_Solutions_Certificate_Authority.pem\nAdding debian:GlobalSign_Root_CA_-_R3.pem\nAdding debian:Verisign_Class_4_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:Security_Communication_RootCA2.pem\nAdding debian:AddTrust_External_Root.pem\nAdding debian:Hellenic_Academic_and_Research_Institutions_RootCA_2011.pem\nAdding debian:COMODO_Certification_Authority.pem\nAdding debian:QuoVadis_Root_CA_2.pem\nAdding debian:GeoTrust_Global_CA.pem\nAdding debian:thawte_Primary_Root_CA_-_G3.pem\nAdding debian:Swisscom_Root_CA_2.pem\nAdding debian:Microsec_e-Szigno_Root_CA.pem\nAdding debian:UTN_USERFirst_Email_Root_CA.pem\nAdding debian:SwissSign_Platinum_CA_-_G2.pem\nAdding debian:Certum_Root_CA.pem\nAdding debian:spi-cacert-2008.pem\nAdding debian:Thawte_Premium_Server_CA.pem\nAdding debian:Global_Chambersign_Root_-_2008.pem\nAdding debian:Root_CA_Generalitat_Valenciana.pem\nAdding debian:GTE_CyberTrust_Global_Root.pem\nAdding debian:SecureTrust_CA.pem\nAdding debian:GeoTrust_Global_CA_2.pem\nAdding debian:NetLock_Qualified_=Class_QA=_Root.pem\nAdding debian:EE_Certification_Centre_Root_CA.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_-_G3.pem\nAdding debian:COMODO_ECC_Certification_Authority.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_2007.pem\nAdding debian:ePKI_Root_Certification_Authority.pem\nAdding debian:E-Guven_Kok_Elektronik_Sertifika_Hizmet_Saglayicisi.pem\nAdding debian:WellsSecure_Public_Root_Certificate_Authority.pem\nAdding debian:TDC_Internet_Root_CA.pem\nAdding debian:Juur-SK.pem\nAdding debian:DigiCert_Assured_ID_Root_CA.pem\nAdding debian:certSIGN_ROOT_CA.pem\nAdding debian:Starfield_Services_Root_Certificate_Authority_-_G2.pem\nAdding debian:EBG_Elektronik_Sertifika_Hizmet_Sağlayıcısı.pem\nAdding debian:AC_Raíz_Certicámara_S.A..pem\nAdding debian:XRamp_Global_CA_Root.pem\nAdding debian:America_Online_Root_Certification_Authority_2.pem\nAdding debian:GeoTrust_Universal_CA.pem\nAdding debian:GlobalSign_Root_CA.pem\nAdding debian:Visa_eCommerce_Root.pem\nAdding debian:Actalis_Authentication_Root_CA.pem\nAdding debian:America_Online_Root_Certification_Authority_1.pem\nAdding debian:NetLock_Express_=Class_C=_Root.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_2.pem\nAdding debian:D-TRUST_Root_Class_3_CA_2_EV_2009.pem\nAdding debian:ApplicationCA_-_Japanese_Government.pem\nAdding debian:QuoVadis_Root_CA_3.pem\nAdding debian:SwissSign_Silver_CA_-_G2.pem\nAdding debian:SecureSign_RootCA11.pem\nAdding debian:Deutsche_Telekom_Root_CA_2.pem\nAdding debian:GeoTrust_Primary_Certification_Authority_-_G3.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:thawte_Primary_Root_CA.pem\nAdding debian:Certigna.pem\nAdding debian:DST_Root_CA_X3.pem\nAdding debian:Baltimore_CyberTrust_Root.pem\nAdding debian:RSA_Root_Certificate_1.pem\nAdding debian:IGC_A.pem\nAdding debian:Trustis_FPS_Root_CA.pem\nAdding debian:PSCProcert.pem\nAdding debian:CA_Disig.pem\nAdding debian:Taiwan_GRCA.pem\nAdding debian:Chambers_of_Commerce_Root_-_2008.pem\nAdding debian:Verisign_Class_2_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:cert-ec2.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_1.pem\nAdding debian:Sonera_Class_2_Root_CA.pem\nAdding debian:DST_ACES_CA_X6.pem\nAdding debian:EC-ACC.pem\nAdding debian:OISTE_WISeKey_Global_Root_GA_CA.pem\nAdding debian:AddTrust_Qualified_Certificates_Root.pem\nAdding debian:Entrust_Root_Certification_Authority.pem\nAdding debian:thawte_Primary_Root_CA_-_G2.pem\nAdding debian:AffirmTrust_Commercial.pem\nAdding debian:Cybertrust_Global_Root.pem\nAdding debian:TDC_OCES_Root_CA.pem\nAdding debian:Sonera_Class_1_Root_CA.pem\nAdding debian:ComSign_Secured_CA.pem\nAdding debian:Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem\nAdding debian:S-TRUST_Authentication_and_Encryption_Root_CA_2005_PN.pem\nAdding debian:Go_Daddy_Class_2_CA.pem\nAdding debian:Comodo_Trusted_Services_root.pem\nAdding debian:Buypass_Class_3_CA_1.pem\nAdding debian:AffirmTrust_Networking.pem\nAdding debian:StartCom_Certification_Authority_2.pem\nAdding debian:Equifax_Secure_Global_eBusiness_CA.pem\nAdding debian:Digital_Signature_Trust_Co._Global_CA_3.pem\nAdding debian:Equifax_Secure_eBusiness_CA_1.pem\nAdding debian:Go_Daddy_Root_Certificate_Authority_-_G2.pem\nAdding debian:AffirmTrust_Premium_ECC.pem\nAdding debian:Microsec_e-Szigno_Root_CA_2009.pem\nAdding debian:Camerfirma_Global_Chambersign_Root.pem\nAdding debian:Comodo_AAA_Services_root.pem\nAdding debian:Swisscom_Root_EV_CA_2.pem\nAdding debian:VeriSign_Universal_Root_Certification_Authority.pem\nAdding debian:VeriSign_Class_3_Public_Primary_Certification_Authority_-_G5.pem\nAdding debian:Equifax_Secure_CA.pem\nAdding debian:Security_Communication_Root_CA.pem\nAdding debian:Entrust.net_Secure_Server_CA.pem\nAdding debian:Buypass_Class_2_Root_CA.pem\nAdding debian:ComSign_CA.pem\nAdding debian:GeoTrust_Primary_Certification_Authority_-_G2.pem\nAdding debian:TÜBİTAK_UEKAE_Kök_Sertifika_Hizmet_Sağlayıcısı_-_Sürüm_3.pem\nAdding debian:Certum_Trusted_Network_CA.pem\nAdding debian:TC_TrustCenter_Class_3_CA_II.pem\nAdding debian:ValiCert_Class_1_VA.pem\nAdding debian:Verisign_Class_1_Public_Primary_Certification_Authority_-_G2.pem\nAdding debian:NetLock_Arany_=Class_Gold=_Főtanúsítvány.pem\nAdding debian:TURKTRUST_Certificate_Services_Provider_Root_2.pem\nAdding debian:Certplus_Class_2_Primary_CA.pem\nAdding debian:Camerfirma_Chambers_of_Commerce_Root.pem\nAdding debian:Security_Communication_EV_RootCA1.pem\nAdding debian:StartCom_Certification_Authority.pem\nAdding debian:Thawte_Server_CA.pem\nAdding debian:RSA_Security_2048_v3.pem\nAdding debian:AddTrust_Low-Value_Services_Root.pem\nAdding debian:Certinomis_-_Autorité_Racine.pem\nAdding debian:Hongkong_Post_Root_CA_1.pem\nAdding debian:ValiCert_Class_2_VA.pem\nAdding debian:Staat_der_Nederlanden_Root_CA_-_G2.pem\nAdding debian:T-TeleSec_GlobalRoot_Class_3.pem\nAdding debian:UTN_DATACorp_SGC_Root_CA.pem\nAdding debian:TC_TrustCenter_Universal_CA_I.pem\nAdding debian:Verisign_Class_3_Public_Primary_Certification_Authority.pem\nAdding debian:TWCA_Root_Certification_Authority.pem\nAdding debian:Buypass_Class_3_Root_CA.pem\nAdding debian:CA_Disig_Root_R1.pem\nAdding debian:Starfield_Class_2_CA.pem\nAdding debian:CNNIC_ROOT.pem\nAdding debian:UTN_USERFirst_Hardware_Root_CA.pem\ndone.\nSetting up libatk-wrapper-java-jni (0.30.4-0ubuntu2) ...\nSetting up openjdk-7-jre (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/jre/bin/policytool to provide /usr/bin/policytool (policytool) in auto mode.\nSetting up openjdk-7-jdk (7u51-2.4.4-0ubuntu0.12.04.2) ...\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/appletviewer to provide /usr/bin/appletviewer (appletviewer) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/extcheck to provide /usr/bin/extcheck (extcheck) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/idlj to provide /usr/bin/idlj (idlj) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jar to provide /usr/bin/jar (jar) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jarsigner to provide /usr/bin/jarsigner (jarsigner) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javac to provide /usr/bin/javac (javac) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javadoc to provide /usr/bin/javadoc (javadoc) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javah to provide /usr/bin/javah (javah) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/javap to provide /usr/bin/javap (javap) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jcmd to provide /usr/bin/jcmd (jcmd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jconsole to provide /usr/bin/jconsole (jconsole) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jdb to provide /usr/bin/jdb (jdb) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jhat to provide /usr/bin/jhat (jhat) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jinfo to provide /usr/bin/jinfo (jinfo) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jmap to provide /usr/bin/jmap (jmap) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jps to provide /usr/bin/jps (jps) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jrunscript to provide /usr/bin/jrunscript (jrunscript) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jsadebugd to provide /usr/bin/jsadebugd (jsadebugd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstack to provide /usr/bin/jstack (jstack) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstat to provide /usr/bin/jstat (jstat) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/jstatd to provide /usr/bin/jstatd (jstatd) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/native2ascii to provide /usr/bin/native2ascii (native2ascii) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/rmic to provide /usr/bin/rmic (rmic) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/schemagen to provide /usr/bin/schemagen (schemagen) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/serialver to provide /usr/bin/serialver (serialver) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/wsgen to provide /usr/bin/wsgen (wsgen) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/wsimport to provide /usr/bin/wsimport (wsimport) in auto mode.\nupdate-alternatives: using /usr/lib/jvm/java-7-openjdk-amd64/bin/xjc to provide /usr/bin/xjc (xjc) in auto mode.\nProcessing triggers for libc-bin ...\nldconfig deferred processing now taking place\naccountsservice\t\t\t\t\tinstall\nacpid\t\t\t\t\t\tinstall\nadduser\t\t\t\t\t\tinstall\napparmor\t\t\t\t\tinstall\napport\t\t\t\t\t\tinstall\napport-symptoms\t\t\t\t\tinstall\napt\t\t\t\t\t\tinstall\napt-transport-https\t\t\t\tinstall\napt-utils\t\t\t\t\tinstall\napt-xapian-index\t\t\t\tinstall\naptitude\t\t\t\t\tinstall\nat\t\t\t\t\t\tinstall\nbase-files\t\t\t\t\tinstall\nbase-passwd\t\t\t\t\tinstall\nbash\t\t\t\t\t\tinstall\nbash-completion\t\t\t\t\tinstall\nbc\t\t\t\t\t\tinstall\nbind9-host\t\t\t\t\tinstall\nbsdmainutils\t\t\t\t\tinstall\nbsdutils\t\t\t\t\tinstall\nbusybox-initramfs\t\t\t\tinstall\nbusybox-static\t\t\t\t\tinstall\nbyobu\t\t\t\t\t\tinstall\nbzip2\t\t\t\t\t\tinstall\nca-certificates\t\t\t\t\tinstall\nca-certificates-java\t\t\t\tinstall\ncloud-init\t\t\t\t\tinstall\ncloud-initramfs-growroot\t\t\tinstall\ncloud-initramfs-rescuevol\t\t\tinstall\ncloud-utils\t\t\t\t\tinstall\ncommand-not-found\t\t\t\tinstall\ncommand-not-found-data\t\t\t\tinstall\nconsole-setup\t\t\t\t\tinstall\nconsolekit\t\t\t\t\tinstall\ncoreutils\t\t\t\t\tinstall\ncpio\t\t\t\t\t\tinstall\ncpp\t\t\t\t\t\tinstall\ncpp-4.6\t\t\t\t\t\tinstall\ncrda\t\t\t\t\t\tinstall\ncron\t\t\t\t\t\tinstall\ncryptsetup-bin\t\t\t\t\tinstall\ncurl\t\t\t\t\t\tinstall\ndash\t\t\t\t\t\tinstall\ndbus\t\t\t\t\t\tinstall\ndbus-x11\t\t\t\t\tinstall\ndconf-gsettings-backend\t\t\t\tinstall\ndconf-service\t\t\t\t\tinstall\ndebconf\t\t\t\t\t\tinstall\ndebconf-i18n\t\t\t\t\tinstall\ndebianutils\t\t\t\t\tinstall\ndiffutils\t\t\t\t\tinstall\ndmidecode\t\t\t\t\tinstall\ndmsetup\t\t\t\t\t\tinstall\ndnsutils\t\t\t\t\tinstall\ndosfstools\t\t\t\t\tinstall\ndpkg\t\t\t\t\t\tinstall\ne2fslibs\t\t\t\t\tinstall\ne2fsprogs\t\t\t\t\tinstall\ned\t\t\t\t\t\tinstall\neject\t\t\t\t\t\tinstall\neuca2ools\t\t\t\t\tinstall\nfile\t\t\t\t\t\tinstall\nfindutils\t\t\t\t\tinstall\nfontconfig\t\t\t\t\tinstall\nfontconfig-config\t\t\t\tinstall\nfonts-ubuntu-font-family-console\t\tinstall\nfriendly-recovery\t\t\t\tinstall\nftp\t\t\t\t\t\tinstall\nfuse\t\t\t\t\t\tinstall\ngcc-4.6-base\t\t\t\t\tinstall\ngconf-service\t\t\t\t\tinstall\ngconf-service-backend\t\t\t\tinstall\ngconf2\t\t\t\t\t\tinstall\ngconf2-common\t\t\t\t\tinstall\ngeoip-database\t\t\t\t\tinstall\ngettext-base\t\t\t\t\tinstall\ngir1.2-glib-2.0\t\t\t\t\tinstall\ngnupg\t\t\t\t\t\tinstall\ngpgv\t\t\t\t\t\tinstall\ngrep\t\t\t\t\t\tinstall\ngroff-base\t\t\t\t\tinstall\ngrub-common\t\t\t\t\tinstall\ngrub-gfxpayload-lists\t\t\t\tinstall\ngrub-legacy-ec2\t\t\t\t\tinstall\ngrub-pc\t\t\t\t\t\tinstall\ngrub-pc-bin\t\t\t\t\tinstall\ngrub2-common\t\t\t\t\tinstall\ngvfs\t\t\t\t\t\tinstall\ngvfs-common\t\t\t\t\tinstall\ngvfs-daemons\t\t\t\t\tinstall\ngvfs-libs\t\t\t\t\tinstall\ngzip\t\t\t\t\t\tinstall\nhdparm\t\t\t\t\t\tinstall\nhicolor-icon-theme\t\t\t\tinstall\nhostname\t\t\t\t\tinstall\nicedtea-7-jre-jamvm\t\t\t\tinstall\nifupdown\t\t\t\t\tinstall\ninfo\t\t\t\t\t\tinstall\ninitramfs-tools\t\t\t\t\tinstall\ninitramfs-tools-bin\t\t\t\tinstall\ninitscripts\t\t\t\t\tinstall\ninsserv\t\t\t\t\t\tinstall\ninstall-info\t\t\t\t\tinstall\niproute\t\t\t\t\t\tinstall\niptables\t\t\t\t\tinstall\niputils-ping\t\t\t\t\tinstall\niputils-tracepath\t\t\t\tinstall\nirqbalance\t\t\t\t\tinstall\nisc-dhcp-client\t\t\t\t\tinstall\nisc-dhcp-common\t\t\t\t\tinstall\niso-codes\t\t\t\t\tinstall\niw\t\t\t\t\t\tinstall\njava-common\t\t\t\t\tinstall\nkbd\t\t\t\t\t\tinstall\nkeyboard-configuration\t\t\t\tinstall\nklibc-utils\t\t\t\t\tinstall\nkrb5-locales\t\t\t\t\tinstall\nlandscape-client\t\t\t\tinstall\nlandscape-common\t\t\t\tinstall\nlanguage-selector-common\t\t\tinstall\nlaptop-detect\t\t\t\t\tinstall\nless\t\t\t\t\t\tinstall\nlibaccountsservice0\t\t\t\tinstall\nlibacl1\t\t\t\t\t\tinstall\nlibapt-inst1.4\t\t\t\t\tinstall\nlibapt-pkg4.12\t\t\t\t\tinstall\nlibasn1-8-heimdal\t\t\t\tinstall\nlibasound2\t\t\t\t\tinstall\nlibasyncns0\t\t\t\t\tinstall\nlibatasmart4\t\t\t\t\tinstall\nlibatk-wrapper-java\t\t\t\tinstall\nlibatk-wrapper-java-jni\t\t\t\tinstall\nlibatk1.0-0\t\t\t\t\tinstall\nlibatk1.0-data\t\t\t\t\tinstall\nlibattr1\t\t\t\t\tinstall\nlibavahi-client3\t\t\t\tinstall\nlibavahi-common-data\t\t\t\tinstall\nlibavahi-common3\t\t\t\tinstall\nlibavahi-glib1\t\t\t\t\tinstall\nlibbind9-80\t\t\t\t\tinstall\nlibblkid1\t\t\t\t\tinstall\nlibbonobo2-0\t\t\t\t\tinstall\nlibbonobo2-common\t\t\t\tinstall\nlibboost-iostreams1.46.1\t\t\tinstall\nlibbsd0\t\t\t\t\t\tinstall\nlibbz2-1.0\t\t\t\t\tinstall\nlibc-bin\t\t\t\t\tinstall\nlibc6\t\t\t\t\t\tinstall\nlibcairo-gobject2\t\t\t\tinstall\nlibcairo2\t\t\t\t\tinstall\nlibcanberra0\t\t\t\t\tinstall\nlibcap-ng0\t\t\t\t\tinstall\nlibck-connector0\t\t\t\tinstall\nlibclass-accessor-perl\t\t\t\tinstall\nlibclass-isa-perl\t\t\t\tinstall\nlibcomerr2\t\t\t\t\tinstall\nlibcryptsetup4\t\t\t\t\tinstall\nlibcups2\t\t\t\t\tinstall\nlibcurl3\t\t\t\t\tinstall\nlibcurl3-gnutls\t\t\t\t\tinstall\nlibcwidget3\t\t\t\t\tinstall\nlibdatrie1\t\t\t\t\tinstall\nlibdb5.1\t\t\t\t\tinstall\nlibdbus-1-3\t\t\t\t\tinstall\nlibdbus-glib-1-2\t\t\t\tinstall\nlibdconf0\t\t\t\t\tinstall\nlibdevmapper-event1.02.1\t\t\tinstall\nlibdevmapper1.02.1\t\t\t\tinstall\nlibdns81\t\t\t\t\tinstall\nlibdrm-intel1\t\t\t\t\tinstall\nlibdrm-nouveau1a\t\t\t\tinstall\nlibdrm-radeon1\t\t\t\t\tinstall\nlibdrm2\t\t\t\t\t\tinstall\nlibedit2\t\t\t\t\tinstall\nlibelf1\t\t\t\t\t\tinstall\nlibept1.4.12\t\t\t\t\tinstall\nlibevent-2.0-5\t\t\t\t\tinstall\nlibexpat1\t\t\t\t\tinstall\nlibffi6\t\t\t\t\t\tinstall\nlibflac8\t\t\t\t\tinstall\nlibfontconfig1\t\t\t\t\tinstall\nlibfontenc1\t\t\t\t\tinstall\nlibfreetype6\t\t\t\t\tinstall\nlibfribidi0\t\t\t\t\tinstall\nlibfuse2\t\t\t\t\tinstall\nlibgc1c2\t\t\t\t\tinstall\nlibgcc1\t\t\t\t\t\tinstall\nlibgconf-2-4\t\t\t\t\tinstall\nlibgconf2-4\t\t\t\t\tinstall\nlibgcrypt11\t\t\t\t\tinstall\nlibgdbm3\t\t\t\t\tinstall\nlibgdk-pixbuf2.0-0\t\t\t\tinstall\nlibgdk-pixbuf2.0-common\t\t\t\tinstall\nlibgdu0\t\t\t\t\t\tinstall\nlibgeoip1\t\t\t\t\tinstall\nlibgif4\t\t\t\t\t\tinstall\nlibgirepository-1.0-1\t\t\t\tinstall\nlibgl1-mesa-dri\t\t\t\t\tinstall\nlibgl1-mesa-glx\t\t\t\t\tinstall\nlibglapi-mesa\t\t\t\t\tinstall\nlibglib2.0-0\t\t\t\t\tinstall\nlibgmp10\t\t\t\t\tinstall\nlibgnome-keyring-common\t\t\t\tinstall\nlibgnome-keyring0\t\t\t\tinstall\nlibgnome2-0\t\t\t\t\tinstall\nlibgnome2-bin\t\t\t\t\tinstall\nlibgnome2-common\t\t\t\tinstall\nlibgnomevfs2-0\t\t\t\t\tinstall\nlibgnomevfs2-common\t\t\t\tinstall\nlibgnutls26\t\t\t\t\tinstall\nlibgpg-error0\t\t\t\t\tinstall\nlibgpm2\t\t\t\t\t\tinstall\nlibgssapi-krb5-2\t\t\t\tinstall\nlibgssapi3-heimdal\t\t\t\tinstall\nlibgtk-3-0\t\t\t\t\tinstall\nlibgtk-3-bin\t\t\t\t\tinstall\nlibgtk-3-common\t\t\t\t\tinstall\nlibgtk2.0-0\t\t\t\t\tinstall\nlibgtk2.0-bin\t\t\t\t\tinstall\nlibgtk2.0-common\t\t\t\tinstall\nlibgudev-1.0-0\t\t\t\t\tinstall\nlibhcrypto4-heimdal\t\t\t\tinstall\nlibheimbase1-heimdal\t\t\t\tinstall\nlibheimntlm0-heimdal\t\t\t\tinstall\nlibhx509-5-heimdal\t\t\t\tinstall\nlibice-dev\t\t\t\t\tinstall\nlibice6\t\t\t\t\t\tinstall\nlibidl-common\t\t\t\t\tinstall\nlibidl0\t\t\t\t\t\tinstall\nlibidn11\t\t\t\t\tinstall\nlibio-string-perl\t\t\t\tinstall\nlibisc83\t\t\t\t\tinstall\nlibisccc80\t\t\t\t\tinstall\nlibisccfg82\t\t\t\t\tinstall\nlibiw30\t\t\t\t\t\tinstall\nlibjasper1\t\t\t\t\tinstall\nlibjpeg-turbo8\t\t\t\t\tinstall\nlibjpeg8\t\t\t\t\tinstall\nlibjs-jquery\t\t\t\t\tinstall\nlibjson0\t\t\t\t\tinstall\nlibk5crypto3\t\t\t\t\tinstall\nlibkeyutils1\t\t\t\t\tinstall\nlibklibc\t\t\t\t\tinstall\nlibkrb5-26-heimdal\t\t\t\tinstall\nlibkrb5-3\t\t\t\t\tinstall\nlibkrb5support0\t\t\t\t\tinstall\nliblcms2-2\t\t\t\t\tinstall\nlibldap-2.4-2\t\t\t\t\tinstall\nlibllvm3.0\t\t\t\t\tinstall\nliblocale-gettext-perl\t\t\t\tinstall\nliblockfile-bin\t\t\t\t\tinstall\nliblockfile1\t\t\t\t\tinstall\nlibltdl7\t\t\t\t\tinstall\nliblvm2app2.2\t\t\t\t\tinstall\nliblwres80\t\t\t\t\tinstall\nliblzma5\t\t\t\t\tinstall\nlibmagic1\t\t\t\t\tinstall\nlibmount1\t\t\t\t\tinstall\nlibmpc2\t\t\t\t\t\tinstall\nlibmpfr4\t\t\t\t\tinstall\nlibncurses5\t\t\t\t\tinstall\nlibncursesw5\t\t\t\t\tinstall\nlibnewt0.52\t\t\t\t\tinstall\nlibnfnetlink0\t\t\t\t\tinstall\nlibnih-dbus1\t\t\t\t\tinstall\nlibnih1\t\t\t\t\t\tinstall\nlibnl-3-200\t\t\t\t\tinstall\nlibnl-genl-3-200\t\t\t\tinstall\nlibnspr4\t\t\t\t\tinstall\nlibnss3\t\t\t\t\t\tinstall\nlibnss3-1d\t\t\t\t\tinstall\nlibogg0\t\t\t\t\t\tinstall\nliborbit2\t\t\t\t\tinstall\nlibp11-kit0\t\t\t\t\tinstall\nlibpam-ck-connector\t\t\t\tinstall\nlibpam-modules\t\t\t\t\tinstall\nlibpam-modules-bin\t\t\t\tinstall\nlibpam-runtime\t\t\t\t\tinstall\nlibpam0g\t\t\t\t\tinstall\nlibpango1.0-0\t\t\t\t\tinstall\nlibparse-debianchangelog-perl\t\t\tinstall\nlibparted0debian1\t\t\t\tinstall\nlibpcap0.8\t\t\t\t\tinstall\nlibpci3\t\t\t\t\t\tinstall\nlibpciaccess0\t\t\t\t\tinstall\nlibpcre3\t\t\t\t\tinstall\nlibpcsclite1\t\t\t\t\tinstall\nlibpipeline1\t\t\t\t\tinstall\nlibpixman-1-0\t\t\t\t\tinstall\nlibplymouth2\t\t\t\t\tinstall\nlibpng12-0\t\t\t\t\tinstall\nlibpolkit-agent-1-0\t\t\t\tinstall\nlibpolkit-backend-1-0\t\t\t\tinstall\nlibpolkit-gobject-1-0\t\t\t\tinstall\nlibpopt0\t\t\t\t\tinstall\nlibpthread-stubs0\t\t\t\tinstall\nlibpthread-stubs0-dev\t\t\t\tinstall\nlibpulse0\t\t\t\t\tinstall\nlibpython2.7\t\t\t\t\tinstall\nlibreadline6\t\t\t\t\tinstall\nlibroken18-heimdal\t\t\t\tinstall\nlibrtmp0\t\t\t\t\tinstall\nlibsasl2-2\t\t\t\t\tinstall\nlibsasl2-modules\t\t\t\tinstall\nlibselinux1\t\t\t\t\tinstall\nlibsgutils2-2\t\t\t\t\tinstall\nlibsigc++-2.0-0c2a\t\t\t\tinstall\nlibslang2\t\t\t\t\tinstall\nlibsm-dev\t\t\t\t\tinstall\nlibsm6\t\t\t\t\t\tinstall\nlibsndfile1\t\t\t\t\tinstall\nlibsqlite3-0\t\t\t\t\tinstall\nlibss2\t\t\t\t\t\tinstall\nlibssl1.0.0\t\t\t\t\tinstall\nlibstdc++6\t\t\t\t\tinstall\nlibsub-name-perl\t\t\t\tinstall\nlibswitch-perl\t\t\t\t\tinstall\nlibtasn1-3\t\t\t\t\tinstall\nlibtdb1\t\t\t\t\t\tinstall\nlibtext-charwidth-perl\t\t\t\tinstall\nlibtext-iconv-perl\t\t\t\tinstall\nlibtext-wrapi18n-perl\t\t\t\tinstall\nlibthai-data\t\t\t\t\tinstall\nlibthai0\t\t\t\t\tinstall\nlibtiff4\t\t\t\t\tinstall\nlibtimedate-perl\t\t\t\tinstall\nlibtinfo5\t\t\t\t\tinstall\nlibudev0\t\t\t\t\tinstall\nlibusb-0.1-4\t\t\t\t\tinstall\nlibusb-1.0-0\t\t\t\t\tinstall\nlibuuid1\t\t\t\t\tinstall\nlibvorbis0a\t\t\t\t\tinstall\nlibvorbisenc2\t\t\t\t\tinstall\nlibvorbisfile3\t\t\t\t\tinstall\nlibwind0-heimdal\t\t\t\tinstall\nlibwrap0\t\t\t\t\tinstall\nlibx11-6\t\t\t\t\tinstall\nlibx11-data\t\t\t\t\tinstall\nlibx11-dev\t\t\t\t\tinstall\nlibx11-doc\t\t\t\t\tinstall\nlibx11-xcb1\t\t\t\t\tinstall\nlibxapian22\t\t\t\t\tinstall\nlibxau-dev\t\t\t\t\tinstall\nlibxau6\t\t\t\t\t\tinstall\nlibxaw7\t\t\t\t\t\tinstall\nlibxcb-glx0\t\t\t\t\tinstall\nlibxcb-render0\t\t\t\t\tinstall\nlibxcb-shape0\t\t\t\t\tinstall\nlibxcb-shm0\t\t\t\t\tinstall\nlibxcb1\t\t\t\t\t\tinstall\nlibxcb1-dev\t\t\t\t\tinstall\nlibxcomposite1\t\t\t\t\tinstall\nlibxcursor1\t\t\t\t\tinstall\nlibxdamage1\t\t\t\t\tinstall\nlibxdmcp-dev\t\t\t\t\tinstall\nlibxdmcp6\t\t\t\t\tinstall\nlibxext6\t\t\t\t\tinstall\nlibxfixes3\t\t\t\t\tinstall\nlibxft2\t\t\t\t\t\tinstall\nlibxi6\t\t\t\t\t\tinstall\nlibxinerama1\t\t\t\t\tinstall\nlibxml2\t\t\t\t\t\tinstall\nlibxmu6\t\t\t\t\t\tinstall\nlibxmuu1\t\t\t\t\tinstall\nlibxpm4\t\t\t\t\t\tinstall\nlibxrandr2\t\t\t\t\tinstall\nlibxrender1\t\t\t\t\tinstall\nlibxt-dev\t\t\t\t\tinstall\nlibxt6\t\t\t\t\t\tinstall\nlibxtst6\t\t\t\t\tinstall\nlibxv1\t\t\t\t\t\tinstall\nlibxxf86dga1\t\t\t\t\tinstall\nlibxxf86vm1\t\t\t\t\tinstall\nlibyaml-0-2\t\t\t\t\tinstall\nlinux-firmware\t\t\t\t\tinstall\nlinux-headers-3.2.0-60\t\t\t\tinstall\nlinux-headers-3.2.0-60-virtual\t\t\tinstall\nlinux-headers-virtual\t\t\t\tinstall\nlinux-image-3.2.0-60-virtual\t\t\tinstall\nlinux-image-virtual\t\t\t\tinstall\nlinux-virtual\t\t\t\t\tinstall\nlocales\t\t\t\t\t\tinstall\nlockfile-progs\t\t\t\t\tinstall\nlogin\t\t\t\t\t\tinstall\nlogrotate\t\t\t\t\tinstall\nlsb-base\t\t\t\t\tinstall\nlsb-release\t\t\t\t\tinstall\nlshw\t\t\t\t\t\tinstall\nlsof\t\t\t\t\t\tinstall\nltrace\t\t\t\t\t\tinstall\nmakedev\t\t\t\t\t\tinstall\nman-db\t\t\t\t\t\tinstall\nmanpages\t\t\t\t\tinstall\nmawk\t\t\t\t\t\tinstall\nmemtest86+\t\t\t\t\tinstall\nmime-support\t\t\t\t\tinstall\nmlocate\t\t\t\t\t\tinstall\nmodule-init-tools\t\t\t\tinstall\nmount\t\t\t\t\t\tinstall\nmountall\t\t\t\t\tinstall\nmtools\t\t\t\t\t\tinstall\nmtr-tiny\t\t\t\t\tinstall\nmultiarch-support\t\t\t\tinstall\nnano\t\t\t\t\t\tinstall\nncurses-base\t\t\t\t\tinstall\nncurses-bin\t\t\t\t\tinstall\nnet-tools\t\t\t\t\tinstall\nnetbase\t\t\t\t\t\tinstall\nnetcat-openbsd\t\t\t\t\tinstall\nntfs-3g\t\t\t\t\t\tinstall\nntpdate\t\t\t\t\t\tinstall\nopenjdk-7-jdk\t\t\t\t\tinstall\nopenjdk-7-jre\t\t\t\t\tinstall\nopenjdk-7-jre-headless\t\t\t\tinstall\nopenssh-client\t\t\t\t\tinstall\nopenssh-server\t\t\t\t\tinstall\nopenssl\t\t\t\t\t\tinstall\nos-prober\t\t\t\t\tinstall\nparted\t\t\t\t\t\tinstall\npasswd\t\t\t\t\t\tinstall\npatch\t\t\t\t\t\tinstall\npciutils\t\t\t\t\tinstall\nperl\t\t\t\t\t\tinstall\nperl-base\t\t\t\t\tinstall\nperl-modules\t\t\t\t\tinstall\nplymouth\t\t\t\t\tinstall\nplymouth-theme-ubuntu-text\t\t\tinstall\npolicykit-1\t\t\t\t\tinstall\npolicykit-1-gnome\t\t\t\tinstall\npopularity-contest\t\t\t\tinstall\npowermgmt-base\t\t\t\t\tinstall\nppp\t\t\t\t\t\tinstall\npppconfig\t\t\t\t\tinstall\npppoeconf\t\t\t\t\tinstall\nprocps\t\t\t\t\t\tinstall\npsmisc\t\t\t\t\t\tinstall\npython\t\t\t\t\t\tinstall\npython-apport\t\t\t\t\tinstall\npython-apt\t\t\t\t\tinstall\npython-apt-common\t\t\t\tinstall\npython-boto\t\t\t\t\tinstall\npython-chardet\t\t\t\t\tinstall\npython-cheetah\t\t\t\t\tinstall\npython-configobj\t\t\t\tinstall\npython-crypto\t\t\t\t\tinstall\npython-dbus\t\t\t\t\tinstall\npython-dbus-dev\t\t\t\t\tinstall\npython-debian\t\t\t\t\tinstall\npython-gdbm\t\t\t\t\tinstall\npython-gi\t\t\t\t\tinstall\npython-gnupginterface\t\t\t\tinstall\npython-httplib2\t\t\t\t\tinstall\npython-keyring\t\t\t\t\tinstall\npython-launchpadlib\t\t\t\tinstall\npython-lazr.restfulclient\t\t\tinstall\npython-lazr.uri\t\t\t\t\tinstall\npython-m2crypto\t\t\t\t\tinstall\npython-minimal\t\t\t\t\tinstall\npython-newt\t\t\t\t\tinstall\npython-oauth\t\t\t\t\tinstall\npython-openssl\t\t\t\t\tinstall\npython-pam\t\t\t\t\tinstall\npython-paramiko\t\t\t\t\tinstall\npython-pkg-resources\t\t\t\tinstall\npython-problem-report\t\t\t\tinstall\npython-pycurl\t\t\t\t\tinstall\npython-serial\t\t\t\t\tinstall\npython-simplejson\t\t\t\tinstall\npython-software-properties\t\t\tinstall\npython-twisted-bin\t\t\t\tinstall\npython-twisted-core\t\t\t\tinstall\npython-twisted-names\t\t\t\tinstall\npython-twisted-web\t\t\t\tinstall\npython-wadllib\t\t\t\t\tinstall\npython-xapian\t\t\t\t\tinstall\npython-yaml\t\t\t\t\tinstall\npython-zope.interface\t\t\t\tinstall\npython2.7\t\t\t\t\tinstall\npython2.7-minimal\t\t\t\tinstall\nreadline-common\t\t\t\t\tinstall\nresolvconf\t\t\t\t\tinstall\nrsync\t\t\t\t\t\tinstall\nrsyslog\t\t\t\t\t\tinstall\nscreen\t\t\t\t\t\tinstall\nsed\t\t\t\t\t\tinstall\nsensible-utils\t\t\t\t\tinstall\nsgml-base\t\t\t\t\tinstall\nshared-mime-info\t\t\t\tinstall\nsound-theme-freedesktop\t\t\t\tinstall\nssh-import-id\t\t\t\t\tinstall\nstrace\t\t\t\t\t\tinstall\nsudo\t\t\t\t\t\tinstall\nsysv-rc\t\t\t\t\t\tinstall\nsysvinit-utils\t\t\t\t\tinstall\ntar\t\t\t\t\t\tinstall\ntasksel\t\t\t\t\t\tinstall\ntasksel-data\t\t\t\t\tinstall\ntcpd\t\t\t\t\t\tinstall\ntcpdump\t\t\t\t\t\tinstall\ntelnet\t\t\t\t\t\tinstall\ntime\t\t\t\t\t\tinstall\ntmux\t\t\t\t\t\tinstall\nttf-dejavu-core\t\t\t\t\tinstall\nttf-dejavu-extra\t\t\t\tinstall\ntzdata\t\t\t\t\t\tinstall\ntzdata-java\t\t\t\t\tinstall\nubuntu-keyring\t\t\t\t\tinstall\nubuntu-minimal\t\t\t\t\tinstall\nubuntu-standard\t\t\t\t\tinstall\nucf\t\t\t\t\t\tinstall\nudev\t\t\t\t\t\tinstall\nudisks\t\t\t\t\t\tinstall\nufw\t\t\t\t\t\tinstall\nunattended-upgrades\t\t\t\tinstall\nupdate-manager-core\t\t\t\tinstall\nupdate-notifier-common\t\t\t\tinstall\nupstart\t\t\t\t\t\tinstall\nureadahead\t\t\t\t\tinstall\nusbutils\t\t\t\t\tinstall\nutil-linux\t\t\t\t\tinstall\nuuid-runtime\t\t\t\t\tinstall\nvim\t\t\t\t\t\tinstall\nvim-common\t\t\t\t\tinstall\nvim-runtime\t\t\t\t\tinstall\nvim-tiny\t\t\t\t\tinstall\nw3m\t\t\t\t\t\tinstall\nwget\t\t\t\t\t\tinstall\nwhiptail\t\t\t\t\tinstall\nwhoopsie\t\t\t\t\tinstall\nwireless-regdb\t\t\t\t\tinstall\nwireless-tools\t\t\t\t\tinstall\nwpasupplicant\t\t\t\t\tinstall\nx11-common\t\t\t\t\tinstall\nx11-utils\t\t\t\t\tinstall\nx11proto-core-dev\t\t\t\tinstall\nx11proto-input-dev\t\t\t\tinstall\nx11proto-kb-dev\t\t\t\t\tinstall\nxauth\t\t\t\t\t\tinstall\nxkb-data\t\t\t\t\tinstall\nxml-core\t\t\t\t\tinstall\nxorg-sgml-doctools\t\t\t\tinstall\nxtrans-dev\t\t\t\t\tinstall\nxz-lzma\t\t\t\t\t\tinstall\nxz-utils\t\t\t\t\tinstall\nzlib1g\t\t\t\t\t\tinstall\n#> [install: install: install-from]: Packages : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho '[install: install: install-from]: Packages...';\n{\n{ debconf-set-selections <<EOF\ndebconf debconf/frontend select noninteractive\ndebconf debconf/frontend seen false\nEOF\n} && enableStart() {\nrm /usr/sbin/policy-rc.d\n} && apt-get -q -y install openjdk-7-jdk+ && dpkg --get-selections\n } || { echo '#> [install: install: install-from]: Packages : FAIL'; exit 1;} >&2 \necho '#> [install: install: install-from]: Packages : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/package, | |
:context "[install: install: install-from]: "} | |
{:exit 0, | |
:out "", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent group hadoop; then /usr/sbin/groupmod hadoop;else /usr/sbin/groupadd --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/group, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out "hadoop:x:999:\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent group hadoop; then /usr/sbin/groupmod hadoop;else /usr/sbin/groupadd --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/group, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out "hadoop:x:999:999::/home/hadoop:/bin/bash\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nif getent passwd hadoop; then /usr/sbin/usermod --group \"hadoop\" --shell \"/bin/bash\" hadoop;else /usr/sbin/useradd -g \"hadoop\" --shell \"/bin/bash\" --create-home --system hadoop;fi\nif getent passwd hadoop; then /usr/sbin/usermod --group \"hadoop\" --shell \"/bin/bash\" hadoop;else /usr/sbin/useradd -g \"hadoop\" --shell \"/bin/bash\" --create-home --system hadoop;fi\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/user, | |
:context "[hadoop-user]: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install: set-environment: system-environment: plan-when: Add java environment to /etc/environment...';\n{\nif ! ( [ -e /etc/environment ] ); then\n{ cat > /etc/environment <<EOFpallet\n# environment file created by pallet\n\nEOFpallet\n }\nfi\npallet_set_env() {\nk=$1\nv=$2\ns=$3\nif ! ( grep \"${s}\" /etc/environment 2>&- ); then\nsed -i -e \"/$${k}=/ d\" /etc/environment && sed -i -e \"$ a \\\\\n${s}\" /etc/environment || exit 1\nfi\n} && vv=\"$(dirname $(dirname $(update-alternatives --query javac | grep Best: | cut -f 2 -d ' ')))\"\npallet_set_env \"JAVA_HOME\" \"${vv}\" \"JAVA_HOME=\\\"${vv}\\\"\"\n } || { echo '#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : FAIL'; exit 1;} >&2 \necho '#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"install: set-environment: system-environment: plan-when: Add java environment to /etc/environment...\n#> install: set-environment: system-environment: plan-when: Add java environment to /etc/environment : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions/exec-script*, | |
:context | |
"install: set-environment: system-environment: plan-when: "} | |
nil | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/dfs/data && chown hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/data && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/data\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/data : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/dfs/name && chown hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/name && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/dfs/name\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/dfs/name : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)//dfs/secondary && chown hadoop $(getent passwd hadoop | cut -d: -f6)//dfs/secondary && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)//dfs/secondary\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)//dfs/secondary : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/mapred/local && chown hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/local && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/local\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/local : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system...\n#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system...';\n{\nmkdir -p $(getent passwd hadoop | cut -d: -f6)/mapred/system && chown hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/system && chgrp hadoop $(getent passwd hadoop | cut -d: -f6)/mapred/system\n } || { echo '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory $(getent passwd hadoop | cut -d: -f6)/mapred/system : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory /var/log/hadoop...\n#> local-dirs: Directory /var/log/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory /var/log/hadoop...';\n{\nmkdir -p /var/log/hadoop && chown hadoop /var/log/hadoop && chgrp hadoop /var/log/hadoop\n } || { echo '#> local-dirs: Directory /var/log/hadoop : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory /var/log/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"local-dirs: Directory /var/run/hadoop...\n#> local-dirs: Directory /var/run/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'local-dirs: Directory /var/run/hadoop...';\n{\nmkdir -p /var/run/hadoop && chown hadoop /var/run/hadoop && chgrp hadoop /var/run/hadoop\n } || { echo '#> local-dirs: Directory /var/run/hadoop : FAIL'; exit 1;} >&2 \necho '#> local-dirs: Directory /var/run/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/directory, | |
:context "local-dirs: "} | |
{:exit 0, | |
:out | |
"hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...\nMD5 sum is 2402b944f6cd21bed1003d9e13107221 .bash_profile\n#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) <<'EOFpallet'\n # base.clj:351\nJAVA_HOME=$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\n # base.clj:352\nPATH=${PATH}:/usr/local/hadoop-0.20.2/bin:/usr/local/hadoop-0.20.2/bin\n\nEOFpallet\n } && contentdiff= && if [ -e $(getent passwd user | cut -d: -f6)/.bash_profile ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\ndiff -u $(getent passwd user | cut -d: -f6)/.bash_profile ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) $(getent passwd user | cut -d: -f6)/.bash_profile && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n}\nfi\n } || { echo '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : FAIL'; exit 1;} >&2 \necho '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out | |
"create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...\n#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...';\n{\nln -f -s /usr/local/hadoop-0.20.2/conf /etc/hadoop\n } || { echo '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : FAIL'; exit 1;} >&2 \necho '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/symbolic-link, | |
:context "create-directories: "} | |
{:exit 0, | |
:out | |
"install-hadoop: install: remote-directory...\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\ninstall-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...\n#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS\ninstall-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\nMD5 sum is 1ddf6e8ff716d17631a46384544293c1 hadoop-0.20.2-cdh3u5.tar.gz\n#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\nUntar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n/home/tom\n#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\n#> install-hadoop: install: remote-directory : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-hadoop: install: remote-directory...';\n{\necho 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown hadoop /usr/local/hadoop-0.20.2 && chgrp hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS' && echo 'install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...';\n{\nmkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\n } || { echo '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS' && echo 'install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))) && if hash curl 2>&-; then\ncurl -o \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --retry 5 --silent --show-error --fail --location \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\nif hash wget 2>&-; then\nwget -O \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --tries 5 --no-verbose --progress=dot:mega \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\necho No download utility available\nexit 1\nfi\nfi && contentdiff= && if [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\ndiff -u ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\nmv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n}\nfi\n } || { echo '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS' && if ! { [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; } || ! { [ -e /usr/local/hadoop-0.20.2/.pallet.directory.md5 ]; } || ! { diff ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5; }; then\necho 'Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\nrdf=$(readlink -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) && cd /usr/local/hadoop-0.20.2 && tar xz --strip-components=1 -f ${rdf} && cd -\n } || { echo '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS'\nif [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; then\ncp ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5\nfi\nfi && echo 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown --recursive hadoop /usr/local/hadoop-0.20.2 && chgrp --recursive hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS'\n } || { echo '#> install-hadoop: install: remote-directory : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-directory : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-directory-action, | |
:context "install-hadoop: install: "} | |
{:exit 0, | |
:out | |
"hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...\n#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) <<'EOFpallet'\n # base.clj:351\nJAVA_HOME=$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\n # base.clj:352\nPATH=${PATH}:/usr/local/hadoop-0.20.2/bin:/usr/local/hadoop-0.20.2/bin\n\nEOFpallet\n } && contentdiff= && if [ -e $(getent passwd user | cut -d: -f6)/.bash_profile ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\ndiff -u $(getent passwd user | cut -d: -f6)/.bash_profile ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/iumAW3lrpxxYHRlKEGwDQQ/$(basename $(getent passwd user | cut -d: -f6)/.bash_profile) $(getent passwd user | cut -d: -f6)/.bash_profile && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 $(getent passwd user | cut -d: -f6)/.bash_profile /var/lib/pallet$(getent passwd user | cut -d: -f6)/.bash_profile.md5\n}\nfi\n } || { echo '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : FAIL'; exit 1;} >&2 \necho '#> hadoop-user: remote-file $(getent passwd user | cut -d: -f6)/.bash_profile : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hadoop-user: "} | |
{:exit 0, | |
:out | |
"create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...\n#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop...';\n{\nln -f -s /usr/local/hadoop-0.20.2/conf /etc/hadoop\n } || { echo '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : FAIL'; exit 1;} >&2 \necho '#> create-directories: Link /usr/local/hadoop-0.20.2/conf as /etc/hadoop : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/symbolic-link, | |
:context "create-directories: "} | |
{:exit 0, | |
:out | |
"install-hadoop: install: remote-directory...\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\ninstall-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...\n#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS\ninstall-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\nUntar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...\n/home/tom\n#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS\ninstall-hadoop: install: Directory /usr/local/hadoop-0.20.2...\n#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS\n#> install-hadoop: install: remote-directory : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'install-hadoop: install: remote-directory...';\n{\necho 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown hadoop /usr/local/hadoop-0.20.2 && chgrp hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS' && echo 'install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))...';\n{\nmkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\n } || { echo '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) : SUCCESS' && echo 'install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))) && if hash curl 2>&-; then\ncurl -o \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --retry 5 --silent --show-error --fail --location \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\nif hash wget 2>&-; then\nwget -O \"${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\" --tries 5 --no-verbose --progress=dot:mega \"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz\"\nelse\necho No download utility available\nexit 1\nfi\nfi && contentdiff= && if [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\ndiff -u ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz))\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ]; then\nmv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AXMJEMFjTI5zsFgMJ7Ky2g/$(basename ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) /var/lib/pallet${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5\n}\nfi\n } || { echo '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-file ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS' && if ! { [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; } || ! { [ -e /usr/local/hadoop-0.20.2/.pallet.directory.md5 ]; } || ! { diff ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5; }; then\necho 'Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)...';\n{\nrdf=$(readlink -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz)) && cd /usr/local/hadoop-0.20.2 && tar xz --strip-components=1 -f ${rdf} && cd -\n } || { echo '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : FAIL'; exit 1;} >&2 \necho '#> Untar ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz) : SUCCESS'\nif [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 ]; then\ncp ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/pcDIIprOxrHmbdkziXSaNA/$(basename /usr/local/hadoop-0.20.2/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz).md5 /usr/local/hadoop-0.20.2/.pallet.directory.md5\nfi\nfi && echo 'install-hadoop: install: Directory /usr/local/hadoop-0.20.2...';\n{\nmkdir -p /usr/local/hadoop-0.20.2 && chown --recursive hadoop /usr/local/hadoop-0.20.2 && chgrp --recursive hadoop /usr/local/hadoop-0.20.2\n } || { echo '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: Directory /usr/local/hadoop-0.20.2 : SUCCESS'\n } || { echo '#> install-hadoop: install: remote-directory : FAIL'; exit 1;} >&2 \necho '#> install-hadoop: install: remote-directory : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-directory-action, | |
:context "install-hadoop: install: "}), | |
:phase :install} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\ncat $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "ssh-key: public-key: "} | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n" | |
{:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:action-symbol pallet.actions/assoc-settings, | |
:context "ssh-key: "}), | |
:phase :collect-ssh-keys} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/hdfs-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/_EBnuFWWUrvF-rm7T4GtmQ/hdfs-site.xml\t2014-04-24 16:57:38.739442001 +0000\n@@ -1,8 +1,15 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>dfs.name.dir</name>\n+ <value>/home/hadoop/dfs/name</value>\n+ </property>\n+ <property>\n+ <name>dfs.data.dir</name>\n+ <value>/home/hadoop/dfs/data</value>\n+ </property>\n+ <property>\n+ <name>dfs.permissions.enabled</name>\n+ <value>true</value>\n+ </property>\n </configuration>\n+\nMD5 sum is 76ffdda75f802f219b1b8ca99f304e35 hdfs-site.xml\n#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>dfs.name.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/dfs/name</value>\n </property>\n <property>\n <name>dfs.data.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/dfs/data</value>\n </property>\n <property>\n <name>dfs.permissions.enabled</name>\n <value>true</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hdfs-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hdfs-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) /usr/local/hadoop-0.20.2/conf/hdfs-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml.md5\n}\nfi\n } || { echo '#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : FAIL'; exit 1;} >&2 \necho '#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hdfs-node-configure21358: settings-config-file: "} | |
{:exit 0, | |
:out | |
"hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n--- /usr/local/hadoop-0.20.2/conf/hadoop-env.sh\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/fkmiNrw9P3Otw0E4qjAoYQ/hadoop-env.sh\t2014-04-24 16:57:40.555442011 +0000\n@@ -1,57 +1,5 @@\n-# Set Hadoop-specific environment variables here.\n-\n-# The only required environment variable is JAVA_HOME. All others are\n-# optional. When running a distributed configuration it is best to\n-# set JAVA_HOME in this file, so that it is correctly defined on\n-# remote nodes.\n-\n-# The java implementation to use. Required.\n-# export JAVA_HOME=/usr/lib/j2sdk1.6-sun\n-\n-# Extra Java CLASSPATH elements. Optional.\n-# export HADOOP_CLASSPATH=\"<extra_entries>:$HADOOP_CLASSPATH\"\n-\n-# The maximum amount of heap to use, in MB. Default is 1000.\n-# export HADOOP_HEAPSIZE=2000\n-\n-# Extra Java runtime options. Empty by default.\n-# if [ \"$HADOOP_OPTS\" == \"\" ]; then export HADOOP_OPTS=-server; else HADOOP_OPTS+=\" -server\"; fi\n-\n-# Command specific options appended to HADOOP_OPTS when specified\n-export HADOOP_NAMENODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_NAMENODE_OPTS\"\n-export HADOOP_SECONDARYNAMENODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_SECONDARYNAMENODE_OPTS\"\n-export HADOOP_DATANODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_DATANODE_OPTS\"\n-export HADOOP_BALANCER_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_BALANCER_OPTS\"\n-export HADOOP_JOBTRACKER_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_JOBTRACKER_OPTS\"\n-# export HADOOP_TASKTRACKER_OPTS=\n-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)\n-# export HADOOP_CLIENT_OPTS\n-\n-# Extra ssh options. Empty by default.\n-# export HADOOP_SSH_OPTS=\"-o ConnectTimeout=1 -o SendEnv=HADOOP_CONF_DIR\"\n-\n-# Where log files are stored. $HADOOP_HOME/logs by default.\n-# export HADOOP_LOG_DIR=${HADOOP_HOME}/logs\n-\n-# File naming remote slave hosts. $HADOOP_HOME/conf/slaves by default.\n-# export HADOOP_SLAVES=${HADOOP_HOME}/conf/slaves\n-\n-# host:path where hadoop code should be rsync'd from. Unset by default.\n-# export HADOOP_MASTER=master:/home/$USER/src/hadoop\n-\n-# Seconds to sleep between slave commands. Unset by default. This\n-# can be useful in large clusters, where, e.g., slave rsyncs can\n-# otherwise arrive faster than the master can service them.\n-# export HADOOP_SLAVE_SLEEP=0.1\n-\n-# The directory where pid files are stored. /tmp by default.\n-# NOTE: this should be set to a directory that can only be written to by \n-# the users that are going to run the hadoop daemons. Otherwise there is\n-# the potential for a symlink attack.\n-# export HADOOP_PID_DIR=/var/hadoop/pids\n-\n-# A string representing this instance of hadoop. $USER by default.\n-# export HADOOP_IDENT_STRING=$USER\n-\n-# The scheduling priority for daemon processes. See 'man nice'.\n-# export HADOOP_NICENESS=10\n+HADOOP_PID_DIR=\"/var/run/hadoop\"\n+HADOOP_LOG_DIR=\"/var/log/hadoop\"\n+HADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\n+HADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\n+JAVA_HOME=\"/usr/lib/jvm/java-7-openjdk-amd64/jre\"\nMD5 sum is 20e6ad108cb8a1c87bbce385afc6f7e3 hadoop-env.sh\n#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hdfs-node-configure21358: env-file: write: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>fs.s3n.awsSecretAccessKey</name>\n <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n </property>\n <property>\n <name>fs.s3.awsSecretAccessKey</name>\n <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n </property>\n <property>\n <name>fs.s3.awsAccessKeyId</name>\n <value>AKIAJSLWRIP4RKOIU5PQ</value>\n </property>\n <property>\n <name>fs.s3n.awsAccessKeyId</name>\n <value>AKIAJSLWRIP4RKOIU5PQ</value>\n </property>\n <property>\n <name>io.file.buffer.size</name>\n <value>65536</value>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.JobSubmissionProtocol</name>\n <value/>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.default</name>\n <value>org.apache.hadoop.net.StandardSocketFactory</value>\n </property>\n <property>\n <name>hadoop.tmp.dir</name>\n <value>/tmp/hadoop</value>\n </property>\n <property>\n <name>io.compression.codecs</name>\n <value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec</value>\n </property>\n <property>\n <name>fs.checkpoint.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)//dfs/secondary</value>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.ClientProtocol</name>\n <value/>\n </property>\n <property>\n <name>fs.trash.interval</name>\n <value>1440</value>\n </property>\n <property>\n <name>fs.default.name</name>\n <value>hdfs://10.183.39.166:8020</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/core-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/core-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) /usr/local/hadoop-0.20.2/conf/core-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml.md5\n}\nfi\n } || { echo '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : FAIL'; exit 1;} >&2 \necho '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/core-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/u1nw_W9hXm5XA_eNl6buSg/core-site.xml\t2014-04-24 16:57:42.227442022 +0000\n@@ -1,8 +1,55 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>fs.s3n.awsSecretAccessKey</name>\n+ <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n+ </property>\n+ <property>\n+ <name>fs.s3.awsSecretAccessKey</name>\n+ <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n+ </property>\n+ <property>\n+ <name>fs.s3.awsAccessKeyId</name>\n+ <value>AKIAJSLWRIP4RKOIU5PQ</value>\n+ </property>\n+ <property>\n+ <name>fs.s3n.awsAccessKeyId</name>\n+ <value>AKIAJSLWRIP4RKOIU5PQ</value>\n+ </property>\n+ <property>\n+ <name>io.file.buffer.size</name>\n+ <value>65536</value>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.JobSubmissionProtocol</name>\n+ <value/>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.default</name>\n+ <value>org.apache.hadoop.net.StandardSocketFactory</value>\n+ </property>\n+ <property>\n+ <name>hadoop.tmp.dir</name>\n+ <value>/tmp/hadoop</value>\n+ </property>\n+ <property>\n+ <name>io.compression.codecs</name>\n+ <value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec</value>\n+ </property>\n+ <property>\n+ <name>fs.checkpoint.dir</name>\n+ <value>/home/hadoop//dfs/secondary</value>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.ClientProtocol</name>\n+ <value/>\n+ </property>\n+ <property>\n+ <name>fs.trash.interval</name>\n+ <value>1440</value>\n+ </property>\n+ <property>\n+ <name>fs.default.name</name>\n+ <value>hdfs://10.183.39.166:8020</value>\n+ </property>\n </configuration>\n+\nMD5 sum is 6ccda35e901dc06c0a38b57b4d26836a core-site.xml\n#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "plan-when: settings-config-file: "} | |
{:exit 0, | |
:out | |
"properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...\n--- /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/2Td6Tx3-756wOphL4_ZADQ/hadoop-metrics.properties\t2014-04-24 16:57:43.875442272 +0000\n@@ -1,64 +1 @@\n-# Configuration of the \"dfs\" context for null\n-dfs.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"dfs\" context for file\n-#dfs.class=org.apache.hadoop.metrics.file.FileContext\n-#dfs.period=10\n-#dfs.fileName=/tmp/dfsmetrics.log\n-\n-# Configuration of the \"dfs\" context for ganglia\n-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)\n-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext31\n-# dfs.period=10\n-# dfs.servers=localhost:8649\n-\n-\n-# Configuration of the \"mapred\" context for null\n-mapred.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"mapred\" context for file\n-#mapred.class=org.apache.hadoop.metrics.file.FileContext\n-#mapred.period=10\n-#mapred.fileName=/tmp/mrmetrics.log\n-\n-# Configuration of the \"mapred\" context for ganglia\n-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)\n-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext31\n-# mapred.period=10\n-# mapred.servers=localhost:8649\n-\n-\n-# Configuration of the \"jvm\" context for null\n-#jvm.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"jvm\" context for file\n-#jvm.class=org.apache.hadoop.metrics.file.FileContext\n-#jvm.period=10\n-#jvm.fileName=/tmp/jvmmetrics.log\n-\n-# Configuration of the \"jvm\" context for ganglia\n-# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# jvm.period=10\n-# jvm.servers=localhost:8649\n-\n-\n-# Configuration of the \"ugi\" context for null\n-ugi.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-\n-# Configuration of the \"fairscheduler\" context for null\n-#fairscheduler.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"fairscheduler\" context for file\n-#fairscheduler.class=org.apache.hadoop.metrics.file.FileContext\n-#fairscheduler.period=10\n-#fairscheduler.fileName=/tmp/fairschedulermetrics.log\n-\n-# Configuration of the \"fairscheduler\" context for ganglia\n-# fairscheduler.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# fairscheduler.period=10\n-# fairscheduler.servers=localhost:8649\n-#\n \nMD5 sum is 68b329da9893e34099c7d8ad5cb9c940 hadoop-metrics.properties\n#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) <<EOFpallet\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n}\nfi\n } || { echo '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : FAIL'; exit 1;} >&2 \necho '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "properties-config-file: "} | |
{:exit 0, | |
:out | |
"env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "env-file: write: "} | |
{:exit 0, | |
:out | |
"14/04/24 16:57:47 INFO namenode.NameNode: STARTUP_MSG: \n/************************************************************\nSTARTUP_MSG: Starting NameNode\nSTARTUP_MSG: host = ip-10-183-39-166.ec2.internal/10.183.39.166\nSTARTUP_MSG: args = [-format]\nSTARTUP_MSG: version = 0.20.2-cdh3u5\nSTARTUP_MSG: build = git://ubuntu-slave02/var/lib/jenkins/workspace/CDH3u5-Full-RC/build/cdh3/hadoop20/0.20.2-cdh3u5/source -r 30233064aaf5f2492bc687d61d72956876102109; compiled by 'jenkins' on Fri Oct 5 17:21:34 PDT 2012\n************************************************************/\nRe-format filesystem in /home/hadoop/dfs/name ? (Y or N) 14/04/24 16:57:47 INFO util.GSet: VM type = 64-bit\n14/04/24 16:57:47 INFO util.GSet: 2% max memory = 17.78 MB\n14/04/24 16:57:47 INFO util.GSet: capacity = 2^21 = 2097152 entries\n14/04/24 16:57:47 INFO util.GSet: recommended=2097152, actual=2097152\n14/04/24 16:57:47 INFO namenode.FSNamesystem: fsOwner=hadoop (auth:SIMPLE)\n14/04/24 16:57:48 INFO namenode.FSNamesystem: supergroup=supergroup\n14/04/24 16:57:48 INFO namenode.FSNamesystem: isPermissionEnabled=true\n14/04/24 16:57:48 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=1000\n14/04/24 16:57:48 INFO namenode.FSNamesystem: isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)\n14/04/24 16:57:48 INFO common.Storage: Image file of size 112 saved in 0 seconds.\n14/04/24 16:57:48 INFO common.Storage: Storage directory /home/hadoop/dfs/name has been successfully formatted.\n14/04/24 16:57:48 INFO namenode.NameNode: SHUTDOWN_MSG: \n/************************************************************\nSHUTDOWN_MSG: Shutting down NameNode at ip-10-183-39-166.ec2.internal/10.183.39.166\n************************************************************/\n/home/tom\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2\nif ! { [ -e /current/VERSION ]; } || null; then echo Y | /usr/local/hadoop-0.20.2/bin/hadoop namenode -format;fi\npwd\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "format-hdfs21405: format-hdfs: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>mapred.local.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/mapred/local</value>\n </property>\n <property>\n <name>mapred.map.tasks.speculative.execution</name>\n <value>true</value>\n </property>\n <property>\n <name>mapred.reduce.parallel.copies</name>\n <value>10</value>\n </property>\n <property>\n <name>mapred.output.compression.type</name>\n <value>BLOCK</value>\n </property>\n <property>\n <name>mapred.reduce.tasks</name>\n <value>5</value>\n </property>\n <property>\n <name>mapred.job.tracker</name>\n <value>10.183.39.166:8021</value>\n </property>\n <property>\n <name>mapred.reduce.tasks.speculative.execution</name>\n <value>false</value>\n </property>\n <property>\n <name>mapred.submit.replication</name>\n <value>10</value>\n </property>\n <property>\n <name>mapred.compress.map.output</name>\n <value>true</value>\n </property>\n <property>\n <name>mapred.system.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/mapred/system</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/mapred-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/mapred-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) /usr/local/hadoop-0.20.2/conf/mapred-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml.md5\n}\nfi\n } || { echo '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : FAIL'; exit 1;} >&2 \necho '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/mapred-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/AD4V84zbBBxuErnt7_Lirg/mapred-site.xml\t2014-04-24 16:57:49.823442045 +0000\n@@ -1,8 +1,43 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>mapred.local.dir</name>\n+ <value>/home/hadoop/mapred/local</value>\n+ </property>\n+ <property>\n+ <name>mapred.map.tasks.speculative.execution</name>\n+ <value>true</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.parallel.copies</name>\n+ <value>10</value>\n+ </property>\n+ <property>\n+ <name>mapred.output.compression.type</name>\n+ <value>BLOCK</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.tasks</name>\n+ <value>5</value>\n+ </property>\n+ <property>\n+ <name>mapred.job.tracker</name>\n+ <value>10.183.39.166:8021</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.tasks.speculative.execution</name>\n+ <value>false</value>\n+ </property>\n+ <property>\n+ <name>mapred.submit.replication</name>\n+ <value>10</value>\n+ </property>\n+ <property>\n+ <name>mapred.compress.map.output</name>\n+ <value>true</value>\n+ </property>\n+ <property>\n+ <name>mapred.system.dir</name>\n+ <value>/home/hadoop/mapred/system</value>\n+ </property>\n </configuration>\n+\nMD5 sum is f8f2b6fac9f97968af6fe32470b2a837 mapred-site.xml\n#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "plan-when: settings-config-file: "} | |
{:exit 0, | |
:out | |
"properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...\n#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) <<EOFpallet\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n}\nfi\n } || { echo '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : FAIL'; exit 1;} >&2 \necho '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "properties-config-file: "} | |
{:exit 0, | |
:out | |
"env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "env-file: write: "} | |
{:exit 0, | |
:out | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\ncat $(getent passwd hadoop | cut -d: -f6)/.ssh/id_rsa.pub\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "ssh-key: public-key: "} | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n" | |
{:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:action-symbol pallet.actions/assoc-settings, | |
:context "ssh-key: "}), | |
:phase :configure} | |
{:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/hdfs-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/_EBnuFWWUrvF-rm7T4GtmQ/hdfs-site.xml\t2014-04-24 16:57:38.644528021 +0000\n@@ -1,8 +1,23 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>dfs.datanode.du.reserved</name>\n+ <value>9</value>\n+ </property>\n+ <property>\n+ <name>dfs.datanode.max.xcievers</name>\n+ <value>4096</value>\n+ </property>\n+ <property>\n+ <name>dfs.name.dir</name>\n+ <value>/home/hadoop/dfs/name</value>\n+ </property>\n+ <property>\n+ <name>dfs.data.dir</name>\n+ <value>/home/hadoop/dfs/data</value>\n+ </property>\n+ <property>\n+ <name>dfs.permissions.enabled</name>\n+ <value>true</value>\n+ </property>\n </configuration>\n+\nMD5 sum is 658fe47c154fe88e67d7dadfefd2e128 hdfs-site.xml\n#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>dfs.datanode.du.reserved</name>\n <value>9</value>\n </property>\n <property>\n <name>dfs.datanode.max.xcievers</name>\n <value>4096</value>\n </property>\n <property>\n <name>dfs.name.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/dfs/name</value>\n </property>\n <property>\n <name>dfs.data.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/dfs/data</value>\n </property>\n <property>\n <name>dfs.permissions.enabled</name>\n <value>true</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hdfs-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hdfs-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/_EBnuFWWUrvF-rm7T4GtmQ/$(basename /usr/local/hadoop-0.20.2/conf/hdfs-site.xml) /usr/local/hadoop-0.20.2/conf/hdfs-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hdfs-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hdfs-site.xml.md5\n}\nfi\n } || { echo '#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : FAIL'; exit 1;} >&2 \necho '#> hdfs-node-configure21358: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hdfs-site.xml : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hdfs-node-configure21358: settings-config-file: "} | |
{:exit 0, | |
:out | |
"hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n--- /usr/local/hadoop-0.20.2/conf/hadoop-env.sh\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/fkmiNrw9P3Otw0E4qjAoYQ/hadoop-env.sh\t2014-04-24 16:57:40.392527936 +0000\n@@ -1,57 +1,5 @@\n-# Set Hadoop-specific environment variables here.\n-\n-# The only required environment variable is JAVA_HOME. All others are\n-# optional. When running a distributed configuration it is best to\n-# set JAVA_HOME in this file, so that it is correctly defined on\n-# remote nodes.\n-\n-# The java implementation to use. Required.\n-# export JAVA_HOME=/usr/lib/j2sdk1.6-sun\n-\n-# Extra Java CLASSPATH elements. Optional.\n-# export HADOOP_CLASSPATH=\"<extra_entries>:$HADOOP_CLASSPATH\"\n-\n-# The maximum amount of heap to use, in MB. Default is 1000.\n-# export HADOOP_HEAPSIZE=2000\n-\n-# Extra Java runtime options. Empty by default.\n-# if [ \"$HADOOP_OPTS\" == \"\" ]; then export HADOOP_OPTS=-server; else HADOOP_OPTS+=\" -server\"; fi\n-\n-# Command specific options appended to HADOOP_OPTS when specified\n-export HADOOP_NAMENODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_NAMENODE_OPTS\"\n-export HADOOP_SECONDARYNAMENODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_SECONDARYNAMENODE_OPTS\"\n-export HADOOP_DATANODE_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_DATANODE_OPTS\"\n-export HADOOP_BALANCER_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_BALANCER_OPTS\"\n-export HADOOP_JOBTRACKER_OPTS=\"-Dcom.sun.management.jmxremote $HADOOP_JOBTRACKER_OPTS\"\n-# export HADOOP_TASKTRACKER_OPTS=\n-# The following applies to multiple commands (fs, dfs, fsck, distcp etc)\n-# export HADOOP_CLIENT_OPTS\n-\n-# Extra ssh options. Empty by default.\n-# export HADOOP_SSH_OPTS=\"-o ConnectTimeout=1 -o SendEnv=HADOOP_CONF_DIR\"\n-\n-# Where log files are stored. $HADOOP_HOME/logs by default.\n-# export HADOOP_LOG_DIR=${HADOOP_HOME}/logs\n-\n-# File naming remote slave hosts. $HADOOP_HOME/conf/slaves by default.\n-# export HADOOP_SLAVES=${HADOOP_HOME}/conf/slaves\n-\n-# host:path where hadoop code should be rsync'd from. Unset by default.\n-# export HADOOP_MASTER=master:/home/$USER/src/hadoop\n-\n-# Seconds to sleep between slave commands. Unset by default. This\n-# can be useful in large clusters, where, e.g., slave rsyncs can\n-# otherwise arrive faster than the master can service them.\n-# export HADOOP_SLAVE_SLEEP=0.1\n-\n-# The directory where pid files are stored. /tmp by default.\n-# NOTE: this should be set to a directory that can only be written to by \n-# the users that are going to run the hadoop daemons. Otherwise there is\n-# the potential for a symlink attack.\n-# export HADOOP_PID_DIR=/var/hadoop/pids\n-\n-# A string representing this instance of hadoop. $USER by default.\n-# export HADOOP_IDENT_STRING=$USER\n-\n-# The scheduling priority for daemon processes. See 'man nice'.\n-# export HADOOP_NICENESS=10\n+HADOOP_PID_DIR=\"/var/run/hadoop\"\n+HADOOP_LOG_DIR=\"/var/log/hadoop\"\n+HADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\n+HADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\n+JAVA_HOME=\"/usr/lib/jvm/java-7-openjdk-amd64/jre\"\nMD5 sum is 20e6ad108cb8a1c87bbce385afc6f7e3 hadoop-env.sh\n#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> hdfs-node-configure21358: env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "hdfs-node-configure21358: env-file: write: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>fs.s3n.awsSecretAccessKey</name>\n <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n </property>\n <property>\n <name>fs.s3.awsSecretAccessKey</name>\n <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n </property>\n <property>\n <name>fs.s3.awsAccessKeyId</name>\n <value>AKIAJSLWRIP4RKOIU5PQ</value>\n </property>\n <property>\n <name>fs.s3n.awsAccessKeyId</name>\n <value>AKIAJSLWRIP4RKOIU5PQ</value>\n </property>\n <property>\n <name>io.file.buffer.size</name>\n <value>65536</value>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.JobSubmissionProtocol</name>\n <value/>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.default</name>\n <value>org.apache.hadoop.net.StandardSocketFactory</value>\n </property>\n <property>\n <name>hadoop.tmp.dir</name>\n <value>/tmp/hadoop</value>\n </property>\n <property>\n <name>io.compression.codecs</name>\n <value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec</value>\n </property>\n <property>\n <name>fs.checkpoint.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)//dfs/secondary</value>\n </property>\n <property>\n <name>hadoop.rpc.socket.factory.class.ClientProtocol</name>\n <value/>\n </property>\n <property>\n <name>fs.trash.interval</name>\n <value>1440</value>\n </property>\n <property>\n <name>fs.default.name</name>\n <value>hdfs://10.183.39.166:8020</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/core-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/core-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/u1nw_W9hXm5XA_eNl6buSg/$(basename /usr/local/hadoop-0.20.2/conf/core-site.xml) /usr/local/hadoop-0.20.2/conf/core-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/core-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/core-site.xml.md5\n}\nfi\n } || { echo '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : FAIL'; exit 1;} >&2 \necho '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/core-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/u1nw_W9hXm5XA_eNl6buSg/core-site.xml\t2014-04-24 16:57:42.052527880 +0000\n@@ -1,8 +1,55 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>fs.s3n.awsSecretAccessKey</name>\n+ <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n+ </property>\n+ <property>\n+ <name>fs.s3.awsSecretAccessKey</name>\n+ <value>Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])</value>\n+ </property>\n+ <property>\n+ <name>fs.s3.awsAccessKeyId</name>\n+ <value>AKIAJSLWRIP4RKOIU5PQ</value>\n+ </property>\n+ <property>\n+ <name>fs.s3n.awsAccessKeyId</name>\n+ <value>AKIAJSLWRIP4RKOIU5PQ</value>\n+ </property>\n+ <property>\n+ <name>io.file.buffer.size</name>\n+ <value>65536</value>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.JobSubmissionProtocol</name>\n+ <value/>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.default</name>\n+ <value>org.apache.hadoop.net.StandardSocketFactory</value>\n+ </property>\n+ <property>\n+ <name>hadoop.tmp.dir</name>\n+ <value>/tmp/hadoop</value>\n+ </property>\n+ <property>\n+ <name>io.compression.codecs</name>\n+ <value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec</value>\n+ </property>\n+ <property>\n+ <name>fs.checkpoint.dir</name>\n+ <value>/home/hadoop//dfs/secondary</value>\n+ </property>\n+ <property>\n+ <name>hadoop.rpc.socket.factory.class.ClientProtocol</name>\n+ <value/>\n+ </property>\n+ <property>\n+ <name>fs.trash.interval</name>\n+ <value>1440</value>\n+ </property>\n+ <property>\n+ <name>fs.default.name</name>\n+ <value>hdfs://10.183.39.166:8020</value>\n+ </property>\n </configuration>\n+\nMD5 sum is 6ccda35e901dc06c0a38b57b4d26836a core-site.xml\n#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/core-site.xml : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "plan-when: settings-config-file: "} | |
{:exit 0, | |
:out | |
"properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...\n--- /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/2Td6Tx3-756wOphL4_ZADQ/hadoop-metrics.properties\t2014-04-24 16:57:43.704528064 +0000\n@@ -1,64 +1 @@\n-# Configuration of the \"dfs\" context for null\n-dfs.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"dfs\" context for file\n-#dfs.class=org.apache.hadoop.metrics.file.FileContext\n-#dfs.period=10\n-#dfs.fileName=/tmp/dfsmetrics.log\n-\n-# Configuration of the \"dfs\" context for ganglia\n-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)\n-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext31\n-# dfs.period=10\n-# dfs.servers=localhost:8649\n-\n-\n-# Configuration of the \"mapred\" context for null\n-mapred.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"mapred\" context for file\n-#mapred.class=org.apache.hadoop.metrics.file.FileContext\n-#mapred.period=10\n-#mapred.fileName=/tmp/mrmetrics.log\n-\n-# Configuration of the \"mapred\" context for ganglia\n-# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter)\n-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext31\n-# mapred.period=10\n-# mapred.servers=localhost:8649\n-\n-\n-# Configuration of the \"jvm\" context for null\n-#jvm.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"jvm\" context for file\n-#jvm.class=org.apache.hadoop.metrics.file.FileContext\n-#jvm.period=10\n-#jvm.fileName=/tmp/jvmmetrics.log\n-\n-# Configuration of the \"jvm\" context for ganglia\n-# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# jvm.period=10\n-# jvm.servers=localhost:8649\n-\n-\n-# Configuration of the \"ugi\" context for null\n-ugi.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-\n-# Configuration of the \"fairscheduler\" context for null\n-#fairscheduler.class=org.apache.hadoop.metrics.spi.NullContext\n-\n-# Configuration of the \"fairscheduler\" context for file\n-#fairscheduler.class=org.apache.hadoop.metrics.file.FileContext\n-#fairscheduler.period=10\n-#fairscheduler.fileName=/tmp/fairschedulermetrics.log\n-\n-# Configuration of the \"fairscheduler\" context for ganglia\n-# fairscheduler.class=org.apache.hadoop.metrics.ganglia.GangliaContext\n-# fairscheduler.period=10\n-# fairscheduler.servers=localhost:8649\n-#\n \nMD5 sum is 68b329da9893e34099c7d8ad5cb9c940 hadoop-metrics.properties\n#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) <<EOFpallet\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n}\nfi\n } || { echo '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : FAIL'; exit 1;} >&2 \necho '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "properties-config-file: "} | |
{:exit 0, | |
:out | |
"env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "env-file: write: "} | |
{:summary nil, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) <<EOFpallet\n<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n <property>\n <name>mapred.local.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/mapred/local</value>\n </property>\n <property>\n <name>tasktracker.http.threads</name>\n <value>46</value>\n </property>\n <property>\n <name>mapred.map.tasks.speculative.execution</name>\n <value>true</value>\n </property>\n <property>\n <name>mapred.reduce.parallel.copies</name>\n <value>10</value>\n </property>\n <property>\n <name>mapred.output.compression.type</name>\n <value>BLOCK</value>\n </property>\n <property>\n <name>mapred.reduce.tasks</name>\n <value>5</value>\n </property>\n <property>\n <name>mapred.tasktracker.reduce.tasks.maximum</name>\n <value>1</value>\n </property>\n <property>\n <name>mapred.job.tracker</name>\n <value>10.183.39.166:8021</value>\n </property>\n <property>\n <name>mapred.reduce.tasks.speculative.execution</name>\n <value>false</value>\n </property>\n <property>\n <name>mapred.submit.replication</name>\n <value>10</value>\n </property>\n <property>\n <name>mapred.compress.map.output</name>\n <value>true</value>\n </property>\n <property>\n <name>mapred.system.dir</name>\n <value>$(getent passwd hadoop | cut -d: -f6)/mapred/system</value>\n </property>\n <property>\n <name>mapred.tasktracker.map.tasks.maximum</name>\n <value>4</value>\n </property>\n</configuration>\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/mapred-site.xml ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/mapred-site.xml ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/AD4V84zbBBxuErnt7_Lirg/$(basename /usr/local/hadoop-0.20.2/conf/mapred-site.xml) /usr/local/hadoop-0.20.2/conf/mapred-site.xml && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/mapred-site.xml /var/lib/pallet/usr/local/hadoop-0.20.2/conf/mapred-site.xml.md5\n}\nfi\n } || { echo '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : FAIL'; exit 1;} >&2 \necho '#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : SUCCESS'\n\nexit $?", | |
:flag-values {}, | |
:flags #{}, | |
:out | |
"plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml...\n--- /usr/local/hadoop-0.20.2/conf/mapred-site.xml\t2012-10-06 00:23:03.000000000 +0000\n+++ /tmp/root/AD4V84zbBBxuErnt7_Lirg/mapred-site.xml\t2014-04-24 16:57:46.932527977 +0000\n@@ -1,8 +1,55 @@\n-<?xml version=\"1.0\"?>\n-<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\n-\n-<!-- Put site-specific property overrides in this file. -->\n-\n-<configuration>\n-\n+<?xml version=\"1.0\" encoding=\"UTF-8\"?><configuration>\n+ <property>\n+ <name>mapred.local.dir</name>\n+ <value>/home/hadoop/mapred/local</value>\n+ </property>\n+ <property>\n+ <name>tasktracker.http.threads</name>\n+ <value>46</value>\n+ </property>\n+ <property>\n+ <name>mapred.map.tasks.speculative.execution</name>\n+ <value>true</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.parallel.copies</name>\n+ <value>10</value>\n+ </property>\n+ <property>\n+ <name>mapred.output.compression.type</name>\n+ <value>BLOCK</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.tasks</name>\n+ <value>5</value>\n+ </property>\n+ <property>\n+ <name>mapred.tasktracker.reduce.tasks.maximum</name>\n+ <value>1</value>\n+ </property>\n+ <property>\n+ <name>mapred.job.tracker</name>\n+ <value>10.183.39.166:8021</value>\n+ </property>\n+ <property>\n+ <name>mapred.reduce.tasks.speculative.execution</name>\n+ <value>false</value>\n+ </property>\n+ <property>\n+ <name>mapred.submit.replication</name>\n+ <value>10</value>\n+ </property>\n+ <property>\n+ <name>mapred.compress.map.output</name>\n+ <value>true</value>\n+ </property>\n+ <property>\n+ <name>mapred.system.dir</name>\n+ <value>/home/hadoop/mapred/system</value>\n+ </property>\n+ <property>\n+ <name>mapred.tasktracker.map.tasks.maximum</name>\n+ <value>4</value>\n+ </property>\n </configuration>\n+\nMD5 sum is dc5b816cf2e7388e84798cdd16dfc431 mapred-site.xml\n#> plan-when: settings-config-file: remote-file /usr/local/hadoop-0.20.2/conf/mapred-site.xml : SUCCESS\n", | |
:exit 0, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "plan-when: settings-config-file: "} | |
{:exit 0, | |
:out | |
"properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...\n#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) <<EOFpallet\n\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) ]; then\nchown hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && chgrp hadoop ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/2Td6Tx3-756wOphL4_ZADQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties) /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties.md5\n}\nfi\n } || { echo '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : FAIL'; exit 1;} >&2 \necho '#> properties-config-file: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-metrics.properties : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "properties-config-file: "} | |
{:exit 0, | |
:out | |
"env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...\n#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh...';\n{\n{\n# -- START pallet implementation function\nverify_md5() {\npath=$1\nstate_path=$2\nmd5_path=$3\nfilediff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\nfilediff=$?\nfi && md5diff= && if [ -e ${state_path} ] && [ -e ${md5_path} ]; then\n( cd $(dirname ${md5_path}) && md5sum --quiet --check $(basename ${md5_path}) )\nmd5diff=$?\nfi && errexit=0 && if [ \"${filediff}\" == \"1\" ]; then\necho Existing file did not match the pallet master copy: FAIL\nerrexit=1\nfi && if [ \"${md5diff}\" == \"1\" ]; then\necho Existing content did not match md5: FAIL\nerrexit=1\nfi && [ \"${errexit}\" == \"0\" ]\n}\n# -- END pallet implementation function\nverify_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n} && mkdir -p $(dirname ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)) && { cat > ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) <<EOFpallet\nHADOOP_PID_DIR=\"/var/run/hadoop\"\nHADOOP_LOG_DIR=\"/var/log/hadoop\"\nHADOOP_SSH_OPTS=\"-o StrictHostKeyChecking=no\"\nHADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true\"\nJAVA_HOME=\"$(dirname $(dirname $(update-alternatives --query java | grep Best: | cut -f 2 -d ' ')))\"\nEOFpallet\n } && contentdiff= && if [ -e /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ] && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\ndiff -u /usr/local/hadoop-0.20.2/conf/hadoop-env.sh ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh)\ncontentdiff=$?\nfi && if ! { [ \"${contentdiff}\" == \"0\" ]; } && [ -e ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) ]; then\nchown root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chgrp root ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && chmod 644 ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) && mv -f ${TMPDIR:-${TEMP:-${TMP:-$(if [ -d /tmp ]; then echo /tmp;else\nif [ -d /var/tmp ]; then echo /var/tmp;else\nif [ -d /use/tmp ]; then echo /usr/tmp;fi\nfi\nfi)}}}/root/fkmiNrw9P3Otw0E4qjAoYQ/$(basename /usr/local/hadoop-0.20.2/conf/hadoop-env.sh) /usr/local/hadoop-0.20.2/conf/hadoop-env.sh && {\n# -- START pallet implementation function\nset_tree_permissions() {\ntemplate_path=$1\nnew_path=$2\ndirpath=$(dirname ${new_path})\ntemplatepath=$(dirname $(if [ -e ${template_path} ]; then readlink -f ${template_path};else echo ${template_path};fi))\nif ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi\ntemplatepath=$(readlink -f ${templatepath})\nmkdir -p ${dirpath} || exit 1\nwhile [ \"/\" != \"${templatepath}\" ] ;do d=${dirpath} && t=${templatepath} && if ! { [ -d ${templatepath} ]; }; then\necho ${templatepath} : Directory does not exist.\nexit 1\nfi && dirpath=$(dirname ${dirpath}) && templatepath=$(dirname ${templatepath}) && if ! { [ \"$(stat -c%G ${t})\" == \"$(stat -c%G ${d})\" ]; }; then chgrp $(stat -c%G ${t}) ${d};fi && if ! { [ \"$(stat -c%a ${t})\" == \"$(stat -c%a ${d})\" ]; }; then chmod $(stat -c%a ${t}) ${d};fi && if ! { [ \"$(stat -c%U ${t})\" == \"$(stat -c%U ${d})\" ]; }; then chown $(stat -c%U ${t}) ${d};fi ; done\n}\n# -- END pallet implementation function\nset_tree_permissions /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_file() {\npath=$1\nstate_path=$2\ncontentdiff= && if [ -e ${path} ] && [ -e ${state_path} ]; then\ndiff -u ${path} ${state_path}\ncontentdiff=$?\nfi && if [ -e ${path} ]; then\nif ! { [ \"${contentdiff}\" == \"0\" ]; }; then cp -f --backup=\"numbered\" ${path} ${state_path};fi\nfi && ls -t ${state_path}.~[0-9]*~ 2> /dev/null | tail -n \"+6\" | xargs \\\n rm --force\n}\n# -- END pallet implementation function\nrecord_file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh\n} && {\n# -- START pallet implementation function\nrecord_md5() {\npath=$1\nmd5_path=$2\n(cp=$(readlink -f ${path}) && cd $(dirname ${cp}) && md5sum $(basename ${cp})\n)>${md5_path} && echo MD5 sum is $(cat ${md5_path})\n}\n# -- END pallet implementation function\nrecord_md5 /usr/local/hadoop-0.20.2/conf/hadoop-env.sh /var/lib/pallet/usr/local/hadoop-0.20.2/conf/hadoop-env.sh.md5\n}\nfi\n } || { echo '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : FAIL'; exit 1;} >&2 \necho '#> env-file: write: remote-file /usr/local/hadoop-0.20.2/conf/hadoop-env.sh : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions-impl/remote-file-action, | |
:context "env-file: write: "}), | |
:phase :configure} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"hadoop-service: start hadoop daemon: Name Node (base.clj:539)...\nstarting namenode, logging to /var/log/hadoop/hadoop-hadoop-namenode-ip-10-183-39-166.out\n#> hadoop-service: start hadoop daemon: Name Node (base.clj:539) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-service: start hadoop daemon: Name Node (base.clj:539)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && if ! ( jps | grep -i namenode ); then /usr/local/hadoop-0.20.2/bin/hadoop-daemon.sh start namenode;fi\n } || { echo '#> hadoop-service: start hadoop daemon: Name Node (base.clj:539) : FAIL'; exit 1;} >&2 \necho '#> hadoop-service: start hadoop daemon: Name Node (base.clj:539) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "hadoop-service: "} | |
{:exit 0, | |
:out | |
"hadoop-service: start hadoop daemon: Job Tracker (base.clj:539)...\nstarting jobtracker, logging to /var/log/hadoop/hadoop-hadoop-jobtracker-ip-10-183-39-166.out\n#> hadoop-service: start hadoop daemon: Job Tracker (base.clj:539) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-service: start hadoop daemon: Job Tracker (base.clj:539)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && if ! ( jps | grep -i jobtracker ); then /usr/local/hadoop-0.20.2/bin/hadoop-daemon.sh start jobtracker;fi\n } || { echo '#> hadoop-service: start hadoop daemon: Job Tracker (base.clj:539) : FAIL'; exit 1;} >&2 \necho '#> hadoop-service: start hadoop daemon: Job Tracker (base.clj:539) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "hadoop-service: "}), | |
:phase :run} | |
{:target | |
{:node | |
hi-hadoop-slave ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.196.170.188 private: 10.186.148.53, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@6bbe9ad5>, | |
:install #< clojure.lang.AFunction$1@2ea74d36>, | |
:settings #< clojure.lang.AFunction$1@69be2c27>, | |
:configure #< clojure.lang.AFunction$1@4444eaad>, | |
:run #< clojure.lang.AFunction$1@88b26df>}, | |
:group-name :hi-hadoop-slave, | |
:network {:inbound-ports (22 50010 50020 50075 8021 50060)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :tasktracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"hadoop-service: start hadoop daemon: Data Node (base.clj:539)...\nstarting datanode, logging to /var/log/hadoop/hadoop-hadoop-datanode-ip-10-186-148-53.out\n#> hadoop-service: start hadoop daemon: Data Node (base.clj:539) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-service: start hadoop daemon: Data Node (base.clj:539)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && if ! ( jps | grep -i datanode ); then /usr/local/hadoop-0.20.2/bin/hadoop-daemon.sh start datanode;fi\n } || { echo '#> hadoop-service: start hadoop daemon: Data Node (base.clj:539) : FAIL'; exit 1;} >&2 \necho '#> hadoop-service: start hadoop daemon: Data Node (base.clj:539) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "hadoop-service: "} | |
{:exit 0, | |
:out | |
"hadoop-service: start hadoop daemon: Task Tracker (base.clj:539)...\nstarting tasktracker, logging to /var/log/hadoop/hadoop-hadoop-tasktracker-ip-10-186-148-53.out\n#> hadoop-service: start hadoop daemon: Task Tracker (base.clj:539) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'hadoop-service: start hadoop daemon: Task Tracker (base.clj:539)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && if ! ( jps | grep -i tasktracker ); then /usr/local/hadoop-0.20.2/bin/hadoop-daemon.sh start tasktracker;fi\n } || { echo '#> hadoop-service: start hadoop daemon: Task Tracker (base.clj:539) : FAIL'; exit 1;} >&2 \necho '#> hadoop-service: start hadoop daemon: Task Tracker (base.clj:539) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "hadoop-service: "}), | |
:phase :run} | |
{:target | |
{:node | |
hi-hadoop-nn ZONE/us-east-1b.REGION/us-east-1.PROVIDER/aws-ec2 null | |
ubuntu paravirtual null ubuntu-us-east-1/images/ubuntu-precise-12.04-amd64-server-20140408.manifest.xml | |
RUNNING | |
public: 54.82.187.163 private: 10.183.39.166, | |
:phases | |
{:pallet/os #< clojure.lang.AFunction$1@28f122f4>, | |
:pallet/os-bs #< clojure.lang.AFunction$1@6f2dd62b>, | |
:bootstrap #< clojure.lang.AFunction$1@724a63d>, | |
:install #< clojure.lang.AFunction$1@771df274>, | |
:settings #< clojure.lang.AFunction$1@4f18dac4>, | |
:run #< clojure.lang.AFunction$1@68c2a26d>, | |
:configure #< clojure.lang.AFunction$1@4235d7d6>, | |
:init | |
#<hadoop$eval21402$fn__21404$fn__21420 palletops.crate.hadoop$eval21402$fn__21404$fn__21420@59ff9fb6>, | |
:collect-ssh-keys | |
#<hadoop$eval21453$fn__21455$fn__21511 palletops.crate.hadoop$eval21453$fn__21455$fn__21511@210f9978>}, | |
:group-name :hi-hadoop-nn, | |
:network {:inbound-ports (22 8020 50070 8021 50030)}, | |
:image | |
{:os-family :ubuntu, :os-version-matches "12.04", :os-64-bit true}, | |
:hardware {:hardware-id "m3.large"}, | |
:count 1, | |
:roles #{:datanode :namenode :jobtracker}, | |
:default-phases [:configure]}, | |
:target-type nil, | |
:plan-state | |
{:host | |
{"us-east-1/i-db664188" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:public-key | |
"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCv6ERRF7Tve0/9OzLa+35wYH0hiCFIhcvQPmvSHbM5Hn2kihx6finbG/assJBy+Pyx7xEkEjwIaCCovkv1dAmUZwVDzNSzjTpxoDHBUGQHL1EkEUtqvPOmntomanDPW9nSBBq7HMKboiHpo6szKiNvDsqHQke4G2JktlHOvga2FehF9KheoXeybn5mCEPsuUUFRP6ZDzt97hLSccBm4PMwvTHIdT8p5auVbPFnReekaGBxz9k7kUVuaUfNubjLijlIBkkPvSgwAoFu+ADGHPjZV1rhGS/L4GB4tFQeW6Riynnofs56GaPYkgVErGrojF/85zCITM5vctBFfnxadvU/ ph_:hi-hadoop-nn_us-east-1/i-db664188_key\n", | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}, | |
"us-east-1/i-da664189" | |
{:hadoop | |
{nil | |
{:mapred.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:group "hadoop", | |
:datanode {:jmx-authenticate false, :jmx-port 3003}, | |
:config.vm.application-ram 5840, | |
:ips | |
{:tasktracker ["10.186.148.53"], | |
:jobtracker ["10.183.39.166"], | |
:namenode ["10.183.39.166"], | |
:datanode ["10.186.148.53" "10.183.39.166"]}, | |
:dist-urls | |
{:cloudera "http://archive.cloudera.com/", | |
:mapr "http://package.mapr.com/releases/", | |
:apache "http://www.apache.org/dist/"}, | |
:dfs.datanode.du.reserved 9N, | |
:jvm.period 10, | |
:fs.s3n.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:fs.s3.awsSecretAccessKey | |
#<SupplierOfInstance Suppliers.ofInstance([identity=AKIAJSLWRIP4RKOIU5PQ, credential=GpbtAYLyGVpLzxJ6UPCrTcpayuO+qup1wYt1siTH])>, | |
:secondary-namenode {:jmx-authenticate false, :jmx-port 3001}, | |
:dfs.datanode.max.xcievers 4096, | |
:dfs.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.local.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/local", | |
:fs.s3.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:dfs.name.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/name", | |
:fs.s3n.awsAccessKeyId "AKIAJSLWRIP4RKOIU5PQ", | |
:config.childtask.mx 1014, | |
:tasktracker {:jmx-authenticate false, :jmx-port 3004}, | |
:config.vm.free-disk 32.0, | |
:config.vm.ram 7680, | |
:home "/usr/local/hadoop-0.20.2", | |
:io.file.buffer.size 65536, | |
:config.tasktracker.mx 384, | |
:ugi.period 10, | |
:fairscheduler.period 10, | |
:fairscheduler.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:env-vars | |
{:HADOOP_PID_DIR "/var/run/hadoop", | |
:HADOOP_LOG_DIR "/var/log/hadoop", | |
:HADOOP_SSH_OPTS "-o StrictHostKeyChecking=no", | |
:HADOOP_OPTS "-Djava.net.preferIPv4Stack=true"}, | |
:ugi.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:namenode-role :namenode, | |
:namenode {:jmx-authenticate false, :jmx-port 3000}, | |
:cloudera-version "3.5", | |
:kernel.fs.file-max 65535, | |
:config.task.mx 5072, | |
:config.vm.disk-size 42.0, | |
:jobtracker {:jmx-authenticate false, :jmx-port 3002}, | |
:etc-config-dir "/etc/hadoop", | |
:dfs.data.dir "$(getent passwd hadoop | cut -d: -f6)/dfs/data", | |
:kernel.vm.overcommit 0, | |
:tasktracker.http.threads 46, | |
:hadoop.rpc.socket.factory.class.JobSubmissionProtocol "", | |
:owner "hadoop", | |
:hadoop.rpc.socket.factory.class.default | |
"org.apache.hadoop.net.StandardSocketFactory", | |
:install-strategy | |
:palletops.crate.hadoop.base/remote-directory, | |
:mapred.map.tasks.speculative.execution true, | |
:version "0.20.2", | |
:mapred.reduce.parallel.copies 10, | |
:jvm.class | |
"org.apache.hadoop.metrics.spi.NoEmitMetricsContext", | |
:mapred.output.compression.type "BLOCK", | |
:config.datanode.mx 384, | |
:data-dir "/tmp/namenode/data", | |
:hadoop.tmp.dir "/tmp/hadoop", | |
:io.compression.codecs | |
"org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec", | |
:config.vm.cores 2.0, | |
:config-dir "/usr/local/hadoop-0.20.2/conf", | |
:fs.checkpoint.dir | |
"$(getent passwd hadoop | cut -d: -f6)//dfs/secondary", | |
:config.vm.free-ram 7316, | |
:hadoop.rpc.socket.factory.class.ClientProtocol "", | |
:dfs.period 10, | |
:log-dir "/var/log/hadoop", | |
:config.os.size 300, | |
:mapred.reduce.tasks 5, | |
:pid-dir "/var/run/hadoop", | |
:user "hadoop", | |
:mapred.tasktracker.reduce.tasks.maximum 1, | |
:hostnames | |
{:tasktracker ["hi-hadoop-slave-da664189"], | |
:jobtracker ["hi-hadoop-nn-db664188"], | |
:namenode ["hi-hadoop-nn-db664188"], | |
:datanode | |
["hi-hadoop-slave-da664189" "hi-hadoop-nn-db664188"]}, | |
:mapred.job.tracker "10.183.39.166:8021", | |
:mapred.reduce.tasks.speculative.execution false, | |
:mapred.submit.replication 10, | |
:fs.trash.interval 1440, | |
:dist :cloudera, | |
:mapred.compress.map.output true, | |
:remote-directory | |
{:url | |
"http://archive.cloudera.com/cdh/3/hadoop-0.20.2-cdh3u5.tar.gz"}, | |
:mapred.system.dir | |
"$(getent passwd hadoop | cut -d: -f6)/mapred/system", | |
:mapred.tasktracker.map.tasks.maximum 4, | |
:mapred.period 10, | |
:kernel.vm.swapiness 0, | |
:dfs.permissions.enabled true, | |
:config.os.cache 1476, | |
:fs.default.name "hdfs://10.183.39.166:8020"}}, | |
:java | |
{nil | |
{:vendor :openjdk, | |
:components #{:jdk}, | |
:version "7", | |
:install-strategy :packages, | |
:packages ("openjdk-7-jdk")}}, | |
:pallet/os | |
{nil | |
{:context "os: ", | |
:action-symbol clj-action13025, | |
:id "Ubuntu", | |
:release "12.04", | |
:mach "x86_64", | |
:rev "3.2.0-60-virtual", | |
:os "Linux", | |
:os-family :ubuntu, | |
:os-version "12.04"}}}}, | |
:pallet.action/action-options {}}, | |
:result | |
({:exit 0, | |
:out | |
"initialise-hdfs: hadoop-exec: hadoop dfsadmin -safemode wait (base.clj:485)...\nSafe mode is OFF\n#> initialise-hdfs: hadoop-exec: hadoop dfsadmin -safemode wait (base.clj:485) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'initialise-hdfs: hadoop-exec: hadoop dfsadmin -safemode wait (base.clj:485)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && /usr/local/hadoop-0.20.2/bin/hadoop dfsadmin -safemode wait\n } || { echo '#> initialise-hdfs: hadoop-exec: hadoop dfsadmin -safemode wait (base.clj:485) : FAIL'; exit 1;} >&2 \necho '#> initialise-hdfs: hadoop-exec: hadoop dfsadmin -safemode wait (base.clj:485) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "initialise-hdfs: hadoop-exec: "} | |
{:exit 0, | |
:out | |
"initialise-hdfs: hadoop-mkdir: hadoop-mkdir /tmp/namenode/data (base.clj:501)...\ntest: File does not exist: /tmp/namenode/data\n#> initialise-hdfs: hadoop-mkdir: hadoop-mkdir /tmp/namenode/data (base.clj:501) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'initialise-hdfs: hadoop-mkdir: hadoop-mkdir /tmp/namenode/data (base.clj:501)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && if ! { /usr/local/hadoop-0.20.2/bin/hadoop fs -test -d /tmp/namenode/data; }; then\n/usr/local/hadoop-0.20.2/bin/hadoop fs -mkdir /tmp/namenode/data\nfi\n } || { echo '#> initialise-hdfs: hadoop-mkdir: hadoop-mkdir /tmp/namenode/data (base.clj:501) : FAIL'; exit 1;} >&2 \necho '#> initialise-hdfs: hadoop-mkdir: hadoop-mkdir /tmp/namenode/data (base.clj:501) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "initialise-hdfs: hadoop-mkdir: "} | |
{:exit 0, | |
:out | |
"initialise-hdfs: hadoop-exec: hadoop fs -chmod +w /tmp/namenode/data (base.clj:485)...\n#> initialise-hdfs: hadoop-exec: hadoop fs -chmod +w /tmp/namenode/data (base.clj:485) : SUCCESS\n", | |
:flags #{}, | |
:flag-values {}, | |
:script | |
"#!/usr/bin/env bash\nset -h\necho 'initialise-hdfs: hadoop-exec: hadoop fs -chmod +w /tmp/namenode/data (base.clj:485)...';\n{\nexport PATH=${PATH}:/usr/local/hadoop-0.20.2 && /usr/local/hadoop-0.20.2/bin/hadoop fs -chmod +w /tmp/namenode/data\n } || { echo '#> initialise-hdfs: hadoop-exec: hadoop fs -chmod +w /tmp/namenode/data (base.clj:485) : FAIL'; exit 1;} >&2 \necho '#> initialise-hdfs: hadoop-exec: hadoop fs -chmod +w /tmp/namenode/data (base.clj:485) : SUCCESS'\n\nexit $?", | |
:summary nil, | |
:action-symbol pallet.actions/exec-script*, | |
:context "initialise-hdfs: hadoop-exec: "}), | |
:phase :init})} | |
nil | |
hadoop-int.hadoop> |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment