Created
November 5, 2012 07:35
-
-
Save zygm0nt/4015815 to your computer and use it in GitHub Desktop.
impala varz
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
<h2>Hadoop Configuration</h2><pre>Configuration: core-default.xml, core-site.xml, mapred-default.xml, mapred-site.xml, yarn-default.xml, yarn-site.xml, hdfs-default.xml, hdfs-site.xml | |
<table><tr><th>Key</th><th>Value</th></tr><tr><td>dfs.datanode.data.dir</td><td>file://${hadoop.tmp.dir}/dfs/data</td</tr><tr><td>dfs.namenode.checkpoint.txns</td><td>40000</td</tr><tr><td>s3.replication</td><td>3</td</tr><tr><td>mapreduce.output.fileoutputformat.compress.type</td><td>RECORD</td</tr><tr><td>mapreduce.jobtracker.jobhistory.lru.cache.size</td><td>5</td</tr><tr><td>dfs.datanode.failed.volumes.tolerated</td><td>0</td</tr><tr><td>hadoop.http.filter.initializers</td><td>org.apache.hadoop.http.lib.StaticUserWebFilter</td</tr><tr><td>mapreduce.cluster.temp.dir</td><td>${hadoop.tmp.dir}/mapred/temp</td</tr><tr><td>mapreduce.reduce.shuffle.memory.limit.percent</td><td>0.25</td</tr><tr><td>yarn.nodemanager.keytab</td><td>/etc/krb5.keytab</td</tr><tr><td>mapreduce.reduce.skip.maxgroups</td><td>0</td</tr><tr><td>dfs.https.server.keystore.resource</td><td>ssl-server.xml</td</tr><tr><td>hadoop.http.authentication.kerberos.keytab</td><td>${user.home}/hadoop.keytab</td</tr><tr><td>yarn.nodemanager.localizer.client.thread-count</td><td>5</td</tr><tr><td>mapreduce.framework.name</td><td>local</td</tr><tr><td>io.file.buffer.size</td><td>4096</td</tr><tr><td>mapreduce.task.tmp.dir</td><td>./tmp</td</tr><tr><td>dfs.namenode.checkpoint.period</td><td>3600</td</tr><tr><td>ipc.client.kill.max</td><td>10</td</tr><tr><td>mapreduce.jobtracker.taskcache.levels</td><td>2</td</tr><tr><td>s3.stream-buffer-size</td><td>4096</td</tr><tr><td>dfs.namenode.secondary.http-address</td><td>0.0.0.0:50090</td</tr><tr><td>dfs.namenode.decommission.interval</td><td>30</td</tr><tr><td>dfs.namenode.http-address</td><td>0.0.0.0:50070</td</tr><tr><td>mapreduce.task.files.preserve.failedtasks</td><td>false</td</tr><tr><td>dfs.encrypt.data.transfer</td><td>false</td</tr><tr><td>dfs.datanode.address</td><td>0.0.0.0:50010</td</tr><tr><td>hadoop.http.authentication.token.validity</td><td>36000</td</tr><tr><td>hadoop.security.group.mapping.ldap.search.filter.group</td><td>(objectClass=group)</td</tr><tr><td>dfs.client.failover.max.attempts</td><td>15</td</tr><tr><td>kfs.client-write-packet-size</td><td>65536</td</tr><tr><td>yarn.admin.acl</td><td>*</td</tr><tr><td>yarn.resourcemanager.application-tokens.master-key-rolling-interval-secs</td><td>86400</td</tr><tr><td>dfs.client.failover.connection.retries.on.timeouts</td><td>0</td</tr><tr><td>mapreduce.map.sort.spill.percent</td><td>0.80</td</tr><tr><td>file.stream-buffer-size</td><td>4096</td</tr><tr><td>dfs.webhdfs.enabled</td><td>false</td</tr><tr><td>ipc.client.connection.maxidletime</td><td>10000</td</tr><tr><td>mapreduce.jobtracker.persist.jobstatus.hours</td><td>1</td</tr><tr><td>yarn.nodemanager.address</td><td>0.0.0.0:0</td</tr><tr><td>dfs.datanode.ipc.address</td><td>0.0.0.0:50020</td</tr><tr><td>yarn.app.mapreduce.am.job.task.listener.thread-count</td><td>30</td</tr><tr><td>dfs.namenode.safemode.extension</td><td>30000</td</tr><tr><td>ha.zookeeper.parent-znode</td><td>/hadoop-ha</td</tr><tr><td>yarn.nodemanager.container-executor.class</td><td>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</td</tr><tr><td>io.skip.checksum.errors</td><td>false</td</tr><tr><td>yarn.resourcemanager.scheduler.client.thread-count</td><td>50</td</tr><tr><td>hadoop.http.authentication.kerberos.principal</td><td>HTTP/_HOST@LOCALHOST</td</tr><tr><td>mapreduce.reduce.log.level</td><td>INFO</td</tr><tr><td>fs.s3.maxRetries</td><td>4</td</tr><tr><td>hadoop.kerberos.kinit.command</td><td>kinit</td</tr><tr><td>yarn.nodemanager.process-kill-wait.ms</td><td>2000</td</tr><tr><td>dfs.namenode.name.dir.restore</td><td>false</td</tr><tr><td>mapreduce.jobtracker.handler.count</td><td>10</td</tr><tr><td>yarn.app.mapreduce.client-am.ipc.max-retries</td><td>1</td</tr><tr><td>dfs.client.use.datanode.hostname</td><td>false</td</tr><tr><td>hadoop.util.hash.type</td><td>murmur</td</tr><tr><td>io.seqfile.lazydecompress</td><td>true</td</tr><tr><td>dfs.datanode.dns.interface</td><td>default</td</tr><tr><td>yarn.nodemanager.disk-health-checker.min-healthy-disks</td><td>0.25</td</tr><tr><td>mapreduce.job.maxtaskfailures.per.tracker</td><td>4</td</tr><tr><td>mapreduce.tasktracker.healthchecker.script.timeout</td><td>600000</td</tr><tr><td>hadoop.security.group.mapping.ldap.search.attr.group.name</td><td>cn</td</tr><tr><td>fs.df.interval</td><td>60000</td</tr><tr><td>dfs.namenode.kerberos.internal.spnego.principal</td><td>${dfs.web.authentication.kerberos.principal}</td</tr><tr><td>mapreduce.jobtracker.address</td><td>local</td</tr><tr><td>mapreduce.tasktracker.tasks.sleeptimebeforesigkill</td><td>5000</td</tr><tr><td>dfs.journalnode.rpc-address</td><td>0.0.0.0:8485</td</tr><tr><td>mapreduce.job.acl-view-job</td><td> </td</tr><tr><td>dfs.client.block.write.replace-datanode-on-failure.policy</td><td>DEFAULT</td</tr><tr><td>dfs.namenode.replication.interval</td><td>3</td</tr><tr><td>dfs.namenode.num.checkpoints.retained</td><td>2</td</tr><tr><td>mapreduce.tasktracker.http.address</td><td>0.0.0.0:50060</td</tr><tr><td>yarn.resourcemanager.scheduler.address</td><td>0.0.0.0:8030</td</tr><tr><td>dfs.datanode.directoryscan.threads</td><td>1</td</tr><tr><td>hadoop.security.group.mapping.ldap.ssl</td><td>false</td</tr><tr><td>mapreduce.task.merge.progress.records</td><td>10000</td</tr><tr><td>dfs.heartbeat.interval</td><td>3</td</tr><tr><td>net.topology.script.number.args</td><td>100</td</tr><tr><td>mapreduce.local.clientfactory.class.name</td><td>org.apache.hadoop.mapred.LocalClientFactory</td</tr><tr><td>dfs.client-write-packet-size</td><td>65536</td</tr><tr><td>io.native.lib.available</td><td>true</td</tr><tr><td>dfs.client.failover.connection.retries</td><td>0</td</tr><tr><td>yarn.nodemanager.disk-health-checker.interval-ms</td><td>120000</td</tr><tr><td>dfs.blocksize</td><td>67108864</td</tr><tr><td>mapreduce.jobhistory.webapp.address</td><td>0.0.0.0:19888</td</tr><tr><td>yarn.resourcemanager.resource-tracker.client.thread-count</td><td>50</td</tr><tr><td>dfs.blockreport.initialDelay</td><td>0</td</tr><tr><td>mapreduce.reduce.markreset.buffer.percent</td><td>0.0</td</tr><tr><td>dfs.ha.tail-edits.period</td><td>60</td</tr><tr><td>mapreduce.admin.user.env</td><td>LD_LIBRARY_PATH=$HADOOP_COMMON_HOME/lib/native</td</tr><tr><td>yarn.nodemanager.health-checker.script.timeout-ms</td><td>1200000</td</tr><tr><td>yarn.resourcemanager.client.thread-count</td><td>50</td</tr><tr><td>file.bytes-per-checksum</td><td>512</td</tr><tr><td>dfs.replication.max</td><td>512</td</tr><tr><td>io.map.index.skip</td><td>0</td</tr><tr><td>mapreduce.task.timeout</td><td>600000</td</tr><tr><td>dfs.datanode.du.reserved</td><td>0</td</tr><tr><td>dfs.support.append</td><td>true</td</tr><tr><td>ftp.blocksize</td><td>67108864</td</tr><tr><td>dfs.client.file-block-storage-locations.num-threads</td><td>10</td</tr><tr><td>yarn.nodemanager.container-manager.thread-count</td><td>20</td</tr><tr><td>ipc.server.listen.queue.size</td><td>128</td</tr><tr><td>yarn.resourcemanager.amliveliness-monitor.interval-ms</td><td>1000</td</tr><tr><td>hadoop.ssl.hostname.verifier</td><td>DEFAULT</td</tr><tr><td>mapreduce.tasktracker.dns.interface</td><td>default</td</tr><tr><td>hadoop.security.group.mapping.ldap.search.attr.member</td><td>member</td</tr><tr><td>mapreduce.tasktracker.outofband.heartbeat</td><td>false</td</tr><tr><td>mapreduce.job.userlog.retain.hours</td><td>24</td</tr><tr><td>yarn.nodemanager.resource.memory-mb</td><td>8192</td</tr><tr><td>dfs.namenode.delegation.token.renew-interval</td><td>86400000</td</tr><tr><td>hadoop.ssl.keystores.factory.class</td><td>org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory</td</tr><tr><td>dfs.datanode.sync.behind.writes</td><td>false</td</tr><tr><td>mapreduce.map.maxattempts</td><td>4</td</tr><tr><td>dfs.datanode.handler.count</td><td>10</td</tr><tr><td>hadoop.ssl.require.client.cert</td><td>false</td</tr><tr><td>ftp.client-write-packet-size</td><td>65536</td</tr><tr><td>ipc.server.tcpnodelay</td><td>false</td</tr><tr><td>mapreduce.task.profile.reduces</td><td>0-2</td</tr><tr><td>hadoop.fuse.connection.timeout</td><td>300</td</tr><tr><td>dfs.permissions.superusergroup</td><td>supergroup</td</tr><tr><td>mapreduce.jobtracker.jobhistory.task.numberprogresssplits</td><td>12</td</tr><tr><td>mapreduce.map.speculative</td><td>true</td</tr><tr><td>fs.ftp.host.port</td><td>21</td</tr><tr><td>dfs.datanode.data.dir.perm</td><td>700</td</tr><tr><td>mapreduce.client.submit.file.replication</td><td>10</td</tr><tr><td>s3native.blocksize</td><td>67108864</td</tr><tr><td>mapreduce.job.ubertask.maxmaps</td><td>9</td</tr><tr><td>dfs.namenode.replication.min</td><td>1</td</tr><tr><td>mapreduce.cluster.acls.enabled</td><td>false</td</tr><tr><td>yarn.nodemanager.localizer.fetch.thread-count</td><td>4</td</tr><tr><td>map.sort.class</td><td>org.apache.hadoop.util.QuickSort</td</tr><tr><td>fs.trash.checkpoint.interval</td><td>0</td</tr><tr><td>dfs.namenode.name.dir</td><td>file://${hadoop.tmp.dir}/dfs/name</td</tr><tr><td>yarn.app.mapreduce.am.staging-dir</td><td>/tmp/hadoop-yarn/staging</td</tr><tr><td>fs.AbstractFileSystem.file.impl</td><td>org.apache.hadoop.fs.local.LocalFs</td</tr><tr><td>yarn.nodemanager.env-whitelist</td><td>JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,YARN_HOME</td</tr><tr><td>dfs.image.compression.codec</td><td>org.apache.hadoop.io.compress.DefaultCodec</td</tr><tr><td>mapreduce.job.reduces</td><td>1</td</tr><tr><td>mapreduce.job.complete.cancel.delegation.tokens</td><td>true</td</tr><tr><td>hadoop.security.group.mapping.ldap.search.filter.user</td><td>(&(objectClass=user)(sAMAccountName={0}))</td</tr><tr><td>yarn.nodemanager.sleep-delay-before-sigkill.ms</td><td>250</td</tr><tr><td>mapreduce.tasktracker.healthchecker.interval</td><td>60000</td</tr><tr><td>mapreduce.jobtracker.heartbeats.in.second</td><td>100</td</tr><tr><td>kfs.bytes-per-checksum</td><td>512</td</tr><tr><td>mapreduce.jobtracker.persist.jobstatus.dir</td><td>/jobtracker/jobsInfo</td</tr><tr><td>dfs.namenode.backup.http-address</td><td>0.0.0.0:50105</td</tr><tr><td>hadoop.rpc.protection</td><td>authentication</td</tr><tr><td>dfs.namenode.https-address</td><td>0.0.0.0:50470</td</tr><tr><td>ftp.stream-buffer-size</td><td>4096</td</tr><tr><td>dfs.ha.log-roll.period</td><td>120</td</tr><tr><td>yarn.resourcemanager.admin.client.thread-count</td><td>1</td</tr><tr><td>yarn.resourcemanager.zookeeper-store.session.timeout-ms</td><td>60000</td</tr><tr><td>file.client-write-packet-size</td><td>65536</td</tr><tr><td>hadoop.http.authentication.simple.anonymous.allowed</td><td>true</td</tr><tr><td>yarn.nodemanager.log.retain-seconds</td><td>10800</td</tr><tr><td>dfs.datanode.drop.cache.behind.reads</td><td>false</td</tr><tr><td>dfs.image.transfer.bandwidthPerSec</td><td>0</td</tr><tr><td>mapreduce.tasktracker.instrumentation</td><td>org.apache.hadoop.mapred.TaskTrackerMetricsInst</td</tr><tr><td>io.mapfile.bloom.size</td><td>1048576</td</tr><tr><td>dfs.ha.fencing.ssh.connect-timeout</td><td>30000</td</tr><tr><td>s3.bytes-per-checksum</td><td>512</td</tr><tr><td>fs.automatic.close</td><td>true</td</tr><tr><td>fs.trash.interval</td><td>0</td</tr><tr><td>hadoop.security.authentication</td><td>simple</td</tr><tr><td>fs.defaultFS</td><td>file:///</td</tr><tr><td>hadoop.ssl.server.conf</td><td>ssl-server.xml</td</tr><tr><td>ipc.client.connect.max.retries</td><td>10</td</tr><tr><td>yarn.resourcemanager.delayed.delegation-token.removal-interval-ms</td><td>30000</td</tr><tr><td>dfs.journalnode.http-address</td><td>0.0.0.0:8480</td</tr><tr><td>mapreduce.jobtracker.taskscheduler</td><td>org.apache.hadoop.mapred.JobQueueTaskScheduler</td</tr><tr><td>mapreduce.job.speculative.speculativecap</td><td>0.1</td</tr><tr><td>yarn.am.liveness-monitor.expiry-interval-ms</td><td>600000</td</tr><tr><td>mapreduce.output.fileoutputformat.compress</td><td>false</td</tr><tr><td>net.topology.node.switch.mapping.impl</td><td>org.apache.hadoop.net.ScriptBasedMapping</td</tr><tr><td>dfs.namenode.replication.considerLoad</td><td>true</td</tr><tr><td>mapreduce.job.counters.max</td><td>120</td</tr><tr><td>yarn.resourcemanager.address</td><td>0.0.0.0:8032</td</tr><tr><td>dfs.client.block.write.retries</td><td>3</td</tr><tr><td>yarn.resourcemanager.nm.liveness-monitor.interval-ms</td><td>1000</td</tr><tr><td>io.map.index.interval</td><td>128</td</tr><tr><td>mapred.child.java.opts</td><td>-Xmx200m</td</tr><tr><td>mapreduce.tasktracker.local.dir.minspacestart</td><td>0</td</tr><tr><td>dfs.client.https.keystore.resource</td><td>ssl-client.xml</td</tr><tr><td>mapreduce.client.progressmonitor.pollinterval</td><td>1000</td</tr><tr><td>mapreduce.jobtracker.tasktracker.maxblacklists</td><td>4</td</tr><tr><td>mapreduce.job.queuename</td><td>default</td</tr><tr><td>yarn.nodemanager.localizer.address</td><td>0.0.0.0:8040</td</tr><tr><td>io.mapfile.bloom.error.rate</td><td>0.005</td</tr><tr><td>mapreduce.job.split.metainfo.maxsize</td><td>10000000</td</tr><tr><td>yarn.nodemanager.delete.thread-count</td><td>4</td</tr><tr><td>ipc.client.tcpnodelay</td><td>false</td</tr><tr><td>yarn.app.mapreduce.am.resource.mb</td><td>1536</td</tr><tr><td>dfs.datanode.dns.nameserver</td><td>default</td</tr><tr><td>mapreduce.map.output.compress.codec</td><td>org.apache.hadoop.io.compress.DefaultCodec</td</tr><tr><td>dfs.namenode.accesstime.precision</td><td>3600000</td</tr><tr><td>mapreduce.map.log.level</td><td>INFO</td</tr><tr><td>io.seqfile.compress.blocksize</td><td>1000000</td</tr><tr><td>mapreduce.tasktracker.taskcontroller</td><td>org.apache.hadoop.mapred.DefaultTaskController</td</tr><tr><td>hadoop.security.groups.cache.secs</td><td>300</td</tr><tr><td>mapreduce.job.end-notification.max.attempts</td><td>5</td</tr><tr><td>yarn.nodemanager.webapp.address</td><td>0.0.0.0:8042</td</tr><tr><td>mapreduce.jobtracker.expire.trackers.interval</td><td>600000</td</tr><tr><td>yarn.resourcemanager.webapp.address</td><td>0.0.0.0:8088</td</tr><tr><td>yarn.nodemanager.health-checker.interval-ms</td><td>600000</td</tr><tr><td>hadoop.security.authorization</td><td>false</td</tr><tr><td>fs.ftp.host</td><td>0.0.0.0</td</tr><tr><td>yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms</td><td>1000</td</tr><tr><td>mapreduce.ifile.readahead</td><td>true</td</tr><tr><td>ha.zookeeper.session-timeout.ms</td><td>5000</td</tr><tr><td>mapreduce.tasktracker.taskmemorymanager.monitoringinterval</td><td>5000</td</tr><tr><td>mapreduce.reduce.shuffle.parallelcopies</td><td>5</td</tr><tr><td>mapreduce.map.skip.maxrecords</td><td>0</td</tr><tr><td>dfs.https.enable</td><td>false</td</tr><tr><td>mapreduce.reduce.shuffle.read.timeout</td><td>180000</td</tr><tr><td>mapreduce.output.fileoutputformat.compress.codec</td><td>org.apache.hadoop.io.compress.DefaultCodec</td</tr><tr><td>mapreduce.jobtracker.instrumentation</td><td>org.apache.hadoop.mapred.JobTrackerMetricsInst</td</tr><tr><td>yarn.nodemanager.remote-app-log-dir-suffix</td><td>logs</td</tr><tr><td>dfs.blockreport.intervalMsec</td><td>21600000</td</tr><tr><td>mapreduce.reduce.speculative</td><td>true</td</tr><tr><td>mapreduce.jobhistory.keytab</td><td>/etc/security/keytab/jhs.service.keytab</td</tr><tr><td>dfs.datanode.balance.bandwidthPerSec</td><td>1048576</td</tr><tr><td>file.blocksize</td><td>67108864</td</tr><tr><td>yarn.resourcemanager.admin.address</td><td>0.0.0.0:8033</td</tr><tr><td>yarn.resourcemanager.resource-tracker.address</td><td>0.0.0.0:8031</td</tr><tr><td>mapreduce.tasktracker.local.dir.minspacekill</td><td>0</td</tr><tr><td>mapreduce.jobtracker.staging.root.dir</td><td>${hadoop.tmp.dir}/mapred/staging</td</tr><tr><td>mapreduce.jobtracker.retiredjobs.cache.size</td><td>1000</td</tr><tr><td>ipc.client.connect.max.retries.on.timeouts</td><td>45</td</tr><tr><td>ha.zookeeper.acl</td><td>world:anyone:rwcda</td</tr><tr><td>yarn.nodemanager.local-dirs</td><td>/tmp/nm-local-dir</td</tr><tr><td>mapreduce.reduce.shuffle.connect.timeout</td><td>180000</td</tr><tr><td>dfs.block.access.key.update.interval</td><td>600</td</tr><tr><td>dfs.block.access.token.lifetime</td><td>600</td</tr><tr><td>mapreduce.job.end-notification.retry.attempts</td><td>5</td</tr><tr><td>mapreduce.jobtracker.system.dir</td><td>${hadoop.tmp.dir}/mapred/system</td</tr><tr><td>yarn.nodemanager.admin-env</td><td>MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX</td</tr><tr><td>mapreduce.jobtracker.jobhistory.block.size</td><td>3145728</td</tr><tr><td>mapreduce.tasktracker.indexcache.mb</td><td>10</td</tr><tr><td>dfs.namenode.checkpoint.check.period</td><td>60</td</tr><tr><td>dfs.client.block.write.replace-datanode-on-failure.enable</td><td>true</td</tr><tr><td>dfs.datanode.directoryscan.interval</td><td>21600</td</tr><tr><td>yarn.nodemanager.container-monitor.interval-ms</td><td>3000</td</tr><tr><td>dfs.default.chunk.view.size</td><td>32768</td</tr><tr><td>mapreduce.job.speculative.slownodethreshold</td><td>1.0</td</tr><tr><td>mapreduce.job.reduce.slowstart.completedmaps</td><td>0.05</td</tr><tr><td>hadoop.security.instrumentation.requires.admin</td><td>false</td</tr><tr><td>dfs.namenode.safemode.min.datanodes</td><td>0</td</tr><tr><td>hadoop.http.authentication.signature.secret.file</td><td>${user.home}/hadoop-http-auth-signature-secret</td</tr><tr><td>mapreduce.reduce.maxattempts</td><td>4</td</tr><tr><td>yarn.nodemanager.localizer.cache.target-size-mb</td><td>10240</td</tr><tr><td>s3native.replication</td><td>3</td</tr><tr><td>dfs.datanode.https.address</td><td>0.0.0.0:50475</td</tr><tr><td>mapreduce.reduce.skip.proc.count.autoincr</td><td>true</td</tr><tr><td>file.replication</td><td>1</td</tr><tr><td>hadoop.hdfs.configuration.version</td><td>1</td</tr><tr><td>ipc.client.idlethreshold</td><td>4000</td</tr><tr><td>hadoop.tmp.dir</td><td>/tmp/hadoop-${user.name}</td</tr><tr><td>mapreduce.jobhistory.address</td><td>0.0.0.0:10020</td</tr><tr><td>mapreduce.jobtracker.restart.recover</td><td>false</td</tr><tr><td>mapreduce.cluster.local.dir</td><td>${hadoop.tmp.dir}/mapred/local</td</tr><tr><td>yarn.ipc.serializer.type</td><td>protocolbuffers</td</tr><tr><td>dfs.namenode.decommission.nodes.per.interval</td><td>5</td</tr><tr><td>dfs.namenode.delegation.key.update-interval</td><td>86400000</td</tr><tr><td>fs.s3.buffer.dir</td><td>${hadoop.tmp.dir}/s3</td</tr><tr><td>dfs.namenode.support.allow.format</td><td>true</td</tr><tr><td>yarn.nodemanager.remote-app-log-dir</td><td>/tmp/logs</td</tr><tr><td>hadoop.work.around.non.threadsafe.getpwuid</td><td>false</td</tr><tr><td>dfs.ha.automatic-failover.enabled</td><td>false</td</tr><tr><td>mapreduce.jobtracker.persist.jobstatus.active</td><td>true</td</tr><tr><td>dfs.namenode.logging.level</td><td>info</td</tr><tr><td>yarn.nodemanager.log-dirs</td><td>/tmp/logs</td</tr><tr><td>dfs.namenode.checkpoint.edits.dir</td><td>${dfs.namenode.checkpoint.dir}</td</tr><tr><td>hadoop.rpc.socket.factory.class.default</td><td>org.apache.hadoop.net.StandardSocketFactory</td</tr><tr><td>yarn.resourcemanager.keytab</td><td>/etc/krb5.keytab</td</tr><tr><td>dfs.datanode.http.address</td><td>0.0.0.0:50075</td</tr><tr><td>mapreduce.task.profile</td><td>false</td</tr><tr><td>dfs.namenode.edits.dir</td><td>${dfs.namenode.name.dir}</td</tr><tr><td>hadoop.fuse.timer.period</td><td>5</td</tr><tr><td>mapreduce.map.skip.proc.count.autoincr</td><td>true</td</tr><tr><td>fs.AbstractFileSystem.viewfs.impl</td><td>org.apache.hadoop.fs.viewfs.ViewFs</td</tr><tr><td>mapreduce.job.speculative.slowtaskthreshold</td><td>1.0</td</tr><tr><td>s3native.stream-buffer-size</td><td>4096</td</tr><tr><td>yarn.nodemanager.delete.debug-delay-sec</td><td>0</td</tr><tr><td>dfs.secondary.namenode.kerberos.internal.spnego.principal</td><td>${dfs.web.authentication.kerberos.principal}</td</tr><tr><td>dfs.namenode.safemode.threshold-pct</td><td>0.999f</td</tr><tr><td>mapreduce.ifile.readahead.bytes</td><td>4194304</td</tr><tr><td>yarn.scheduler.maximum-allocation-mb</td><td>10240</td</tr><tr><td>s3native.bytes-per-checksum</td><td>512</td</tr><tr><td>mapreduce.job.committer.setup.cleanup.needed</td><td>true</td</tr><tr><td>kfs.replication</td><td>3</td</tr><tr><td>yarn.nodemanager.log-aggregation.compression-type</td><td>none</td</tr><tr><td>hadoop.http.authentication.type</td><td>simple</td</tr><tr><td>dfs.client.failover.sleep.base.millis</td><td>500</td</tr><tr><td>yarn.nodemanager.heartbeat.interval-ms</td><td>1000</td</tr><tr><td>hadoop.jetty.logs.serve.aliases</td><td>true</td</tr><tr><td>mapreduce.reduce.shuffle.input.buffer.percent</td><td>0.70</td</tr><tr><td>dfs.datanode.max.transfer.threads</td><td>4096</td</tr><tr><td>mapreduce.task.io.sort.mb</td><td>100</td</tr><tr><td>mapreduce.reduce.merge.inmem.threshold</td><td>1000</td</tr><tr><td>dfs.namenode.handler.count</td><td>10</td</tr><tr><td>hadoop.ssl.client.conf</td><td>ssl-client.xml</td</tr><tr><td>yarn.resourcemanager.container.liveness-monitor.interval-ms</td><td>600000</td</tr><tr><td>mapreduce.client.completion.pollinterval</td><td>5000</td</tr><tr><td>yarn.nodemanager.vmem-pmem-ratio</td><td>2.1</td</tr><tr><td>yarn.app.mapreduce.client.max-retries</td><td>3</td</tr><tr><td>hadoop.ssl.enabled</td><td>false</td</tr><tr><td>fs.AbstractFileSystem.hdfs.impl</td><td>org.apache.hadoop.fs.Hdfs</td</tr><tr><td>mapreduce.tasktracker.reduce.tasks.maximum</td><td>2</td</tr><tr><td>mapreduce.reduce.input.buffer.percent</td><td>0.0</td</tr><tr><td>kfs.stream-buffer-size</td><td>4096</td</tr><tr><td>dfs.namenode.invalidate.work.pct.per.iteration</td><td>0.32f</td</tr><tr><td>dfs.bytes-per-checksum</td><td>512</td</tr><tr><td>dfs.replication</td><td>3</td</tr><tr><td>mapreduce.shuffle.ssl.file.buffer.size</td><td>65536</td</tr><tr><td>dfs.permissions.enabled</td><td>true</td</tr><tr><td>mapreduce.jobtracker.maxtasks.perjob</td><td>-1</td</tr><tr><td>dfs.datanode.use.datanode.hostname</td><td>false</td</tr><tr><td>mapreduce.task.userlog.limit.kb</td><td>0</td</tr><tr><td>dfs.namenode.fs-limits.max-directory-items</td><td>0</td</tr><tr><td>s3.client-write-packet-size</td><td>65536</td</tr><tr><td>dfs.client.failover.sleep.max.millis</td><td>15000</td</tr><tr><td>mapreduce.job.maps</td><td>2</td</tr><tr><td>dfs.namenode.fs-limits.max-component-length</td><td>0</td</tr><tr><td>s3.blocksize</td><td>67108864</td</tr><tr><td>mapreduce.map.output.compress</td><td>false</td</tr><tr><td>kfs.blocksize</td><td>67108864</td</tr><tr><td>dfs.namenode.edits.journal-plugin.qjournal</td><td>org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager</td</tr><tr><td>dfs.client.https.need-auth</td><td>false</td</tr><tr><td>yarn.scheduler.minimum-allocation-mb</td><td>128</td</tr><tr><td>ftp.replication</td><td>3</td</tr><tr><td>mapreduce.input.fileinputformat.split.minsize</td><td>0</td</tr><tr><td>fs.s3n.block.size</td><td>67108864</td</tr><tr><td>yarn.ipc.rpc.class</td><td>org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC</td</tr><tr><td>dfs.namenode.num.extra.edits.retained</td><td>1000000</td</tr><tr><td>hadoop.http.staticuser.user</td><td>dr.who</td</tr><tr><td>yarn.nodemanager.localizer.cache.cleanup.interval-ms</td><td>600000</td</tr><tr><td>mapreduce.job.jvm.numtasks</td><td>1</td</tr><tr><td>mapreduce.task.profile.maps</td><td>0-2</td</tr><tr><td>mapreduce.shuffle.port</td><td>8080</td</tr><tr><td>mapreduce.jobtracker.http.address</td><td>0.0.0.0:50030</td</tr><tr><td>mapreduce.reduce.shuffle.merge.percent</td><td>0.66</td</tr><tr><td>mapreduce.task.skip.start.attempts</td><td>2</td</tr><tr><td>mapreduce.task.io.sort.factor</td><td>10</td</tr><tr><td>dfs.namenode.checkpoint.dir</td><td>file://${hadoop.tmp.dir}/dfs/namesecondary</td</tr><tr><td>tfile.fs.input.buffer.size</td><td>262144</td</tr><tr><td>fs.s3.block.size</td><td>67108864</td</tr><tr><td>tfile.io.chunk.size</td><td>1048576</td</tr><tr><td>io.serializations</td><td>org.apache.hadoop.io.serializer.WritableSerialization,org.apache.hadoop.io.serializer.avro.AvroSpecificSerialization,org.apache.hadoop.io.serializer.avro.AvroReflectSerialization</td</tr><tr><td>yarn.resourcemanager.max-completed-applications</td><td>10000</td</tr><tr><td>mapreduce.jobhistory.principal</td><td>jhs/[email protected]</td</tr><tr><td>mapreduce.job.end-notification.retry.interval</td><td>1</td</tr><tr><td>dfs.namenode.backup.address</td><td>0.0.0.0:50100</td</tr><tr><td>dfs.block.access.token.enable</td><td>false</td</tr><tr><td>io.seqfile.sorter.recordlimit</td><td>1000000</td</tr><tr><td>s3native.client-write-packet-size</td><td>65536</td</tr><tr><td>ftp.bytes-per-checksum</td><td>512</td</tr><tr><td>hadoop.security.group.mapping</td><td>org.apache.hadoop.security.ShellBasedUnixGroupsMapping</td</tr><tr><td>dfs.client.file-block-storage-locations.timeout</td><td>60</td</tr><tr><td>mapreduce.job.end-notification.max.retry.interval</td><td>5</td</tr><tr><td>yarn.acl.enable</td><td>true</td</tr><tr><td>yarn.nm.liveness-monitor.expiry-interval-ms</td><td>600000</td</tr><tr><td>mapreduce.tasktracker.map.tasks.maximum</td><td>2</td</tr><tr><td>dfs.namenode.max.objects</td><td>0</td</tr><tr><td>dfs.namenode.delegation.token.max-lifetime</td><td>604800000</td</tr><tr><td>mapreduce.job.hdfs-servers</td><td>${fs.defaultFS}</td</tr><tr><td>yarn.application.classpath</td><td>$HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/share/hadoop/common/*,$HADOOP_COMMON_HOME/share/hadoop/common/lib/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,$YARN_HOME/share/hadoop/yarn/*,$YARN_HOME/share/hadoop/yarn/lib/*,$YARN_HOME/share/hadoop/mapreduce/*,$YARN_HOME/share/hadoop/mapreduce/lib/*</td</tr><tr><td>dfs.datanode.hdfs-blocks-metadata.enabled</td><td>true</td</tr><tr><td>yarn.nodemanager.aux-services.mapreduce.shuffle.class</td><td>org.apache.hadoop.mapred.ShuffleHandler</td</tr><tr><td>mapreduce.tasktracker.dns.nameserver</td><td>default</td</tr><tr><td>dfs.datanode.readahead.bytes</td><td>4193404</td</tr><tr><td>mapreduce.job.ubertask.maxreduces</td><td>1</td</tr><tr><td>dfs.image.compress</td><td>false</td</tr><tr><td>mapreduce.shuffle.ssl.enabled</td><td>false</td</tr><tr><td>yarn.log-aggregation-enable</td><td>false</td</tr><tr><td>mapreduce.tasktracker.report.address</td><td>127.0.0.1:0</td</tr><tr><td>mapreduce.tasktracker.http.threads</td><td>40</td</tr><tr><td>dfs.stream-buffer-size</td><td>4096</td</tr><tr><td>tfile.fs.output.buffer.size</td><td>262144</td</tr><tr><td>yarn.resourcemanager.am.max-retries</td><td>1</td</tr><tr><td>dfs.datanode.drop.cache.behind.writes</td><td>false</td</tr><tr><td>mapreduce.job.ubertask.enable</td><td>false</td</tr><tr><td>hadoop.common.configuration.version</td><td>0.23.0</td</tr><tr><td>dfs.namenode.replication.work.multiplier.per.iteration</td><td>2</td</tr><tr><td>mapreduce.job.acl-modify-job</td><td> </td</tr><tr><td>io.seqfile.local.dir</td><td>${hadoop.tmp.dir}/io/local</td</tr><tr><td>fs.s3.sleepTimeSeconds</td><td>10</td</tr><tr><td>mapreduce.client.output.filter</td><td>FAILED</td</tr></table></pre><h2>Command-line Flags</h2><pre>--dump_ir=false | |
--module_output= | |
--be_port=22000 | |
--classpath= | |
--hostname=centos3 | |
--ipaddress=127.0.0.1 | |
--keytab_file= | |
--planservice_host=localhost | |
--planservice_port=20000 | |
--principal= | |
--randomize_scan_ranges=false | |
--port=20001 | |
--num_disks=0 | |
--num_threads_per_disk=1 | |
--read_size=8388608 | |
--enable_webserver=true | |
--use_statestore=true | |
--nn=localhost | |
--nn_port=20500 | |
--serialize_batch=false | |
--status_report_interval=5 | |
--be_service_threads=64 | |
--default_num_nodes=1 | |
--fe_port=21000 | |
--fe_service_threads=64 | |
--load_catalog_at_startup=false | |
--use_planservice=false | |
--statestore_subscriber_timeout_seconds=10 | |
--statestore_max_missed_heartbeats=5 | |
--statestore_num_server_worker_threads=4 | |
--statestore_suspect_heartbeats=2 | |
--state_store_host=localhost | |
--state_store_port=24000 | |
--state_store_subscriber_port=23000 | |
--kerberos_reinit_interval=60 | |
--kerberos_ticket_life=0 | |
--sasl_path=/usr/lib/sasl2:/usr/lib64/sasl2:/usr/local/lib/sasl2 | |
--web_log_bytes=1048576 | |
--log_filename=impalad | |
--rpc_cnxn_attempts=10 | |
--rpc_cnxn_retry_interval_ms=2000 | |
--webserver_interface= | |
--webserver_port=25000 | |
--flagfile= | |
--fromenv= | |
--tryfromenv= | |
--undefok= | |
--tab_completion_columns=80 | |
--tab_completion_word= | |
--help=false | |
--helpfull=false | |
--helpmatch= | |
--helpon= | |
--helppackage=false | |
--helpshort=false | |
--helpxml=false | |
--version=false | |
</pre> |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment