import java.util.Iterator;
import java.util.Map;
import java.util.Map.Entry;
import org.apache.hadoop.conf.Configuration;
public class HadoopInfo {
public static void main(String[] args) throws Exception {
Configuration conf = new Configuration();
Iterator< Entry< String, String > > entries = conf.iterator();
System.out.println("< table border=\"1\" width=\"760\" style=\"word-break:break-all;\" >" +
"< caption>Hadoop defaul Configruation keys and values< /caption > " +
"< tr >< th >Key< /th >< th >Value< /th >< /tr >");
while(entries.hasNext()){
Map.Entry< String, String> en = (Map.Entry < String, String >)entries.next();
System.out.println("< tr >< td width=\"350\"> " + en.getKey() + " < /td >< td >" + en.getValue() + "< /td >< /tr >");
}
System.out.println("< / table >");
}
}
Sample output for a fresh standalone Hadoop instance.
| Key | Value |
|---|---|
| io.seqfile.compress.blocksize | 1000000 |
| hadoop.http.authentication.signature.secret.file | ${user.home}/hadoop-http-auth-signature-secret |
| io.skip.checksum.errors | false |
| fs.checkpoint.size | 67108864 |
| hadoop.http.authentication.kerberos.principal | HTTP/localhost@LOCALHOST |
| fs.s3n.impl | org.apache.hadoop.fs.s3native.NativeS3FileSystem |
| fs.s3.maxRetries | 4 |
| webinterface.private.actions | false |
| hadoop.http.authentication.simple.anonymous.allowed | true |
| fs.s3.impl | org.apache.hadoop.fs.s3.S3FileSystem |
| hadoop.native.lib | true |
| fs.checkpoint.edits.dir | ${fs.checkpoint.dir} |
| ipc.server.listen.queue.size | 128 |
| fs.default.name | file:/// |
| hadoop.http.authentication.kerberos.keytab | ${user.home}/hadoop.keytab |
| ipc.client.idlethreshold | 4000 |
| hadoop.tmp.dir | /tmp/hadoop-${user.name} |
| fs.hsftp.impl | org.apache.hadoop.hdfs.HsftpFileSystem |
| fs.checkpoint.dir | ${hadoop.tmp.dir}/dfs/namesecondary |
| fs.s3.block.size | 67108864 |
| hadoop.security.authorization | false |
| io.serializations | org.apache.hadoop.io.serializer.WritableSerialization |
| hadoop.util.hash.type | murmur |
| io.seqfile.lazydecompress | true |
| io.file.buffer.size | 4096 |
| io.mapfile.bloom.size | 1048576 |
| fs.s3.buffer.dir | ${hadoop.tmp.dir}/s3 |
| hadoop.logfile.size | 10000000 |
| fs.webhdfs.impl | org.apache.hadoop.hdfs.web.WebHdfsFileSystem |
| ipc.client.kill.max | 10 |
| io.compression.codecs | org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec,org.apache.hadoop.io.compress.SnappyCodec |
| topology.script.number.args | 100 |
| fs.har.impl | org.apache.hadoop.fs.HarFileSystem |
| io.seqfile.sorter.recordlimit | 1000000 |
| fs.trash.interval | 0 |
| hadoop.security.authentication | simple |
| local.cache.size | 10737418240 |
| hadoop.security.group.mapping | org.apache.hadoop.security.ShellBasedUnixGroupsMapping |
| ipc.server.tcpnodelay | false |
| hadoop.security.token.service.use_ip | true |
| fs.ramfs.impl | org.apache.hadoop.fs.InMemoryFileSystem |
| ipc.client.connect.max.retries | 10 |
| hadoop.rpc.socket.factory.class.default | org.apache.hadoop.net.StandardSocketFactory |
| fs.kfs.impl | org.apache.hadoop.fs.kfs.KosmosFileSystem |
| fs.checkpoint.period | 3600 |
| topology.node.switch.mapping.impl | org.apache.hadoop.net.ScriptBasedMapping |
| hadoop.http.authentication.token.validity | 36000 |
| hadoop.security.use-weak-http-crypto | false |
| hadoop.logfile.count | 10 |
| hadoop.security.uid.cache.secs | 14400 |
| fs.ftp.impl | org.apache.hadoop.fs.ftp.FTPFileSystem |
| fs.file.impl | org.apache.hadoop.fs.LocalFileSystem |
| fs.hdfs.impl | org.apache.hadoop.hdfs.DistributedFileSystem |
| ipc.client.connection.maxidletime | 10000 |
| io.mapfile.bloom.error.rate | 0.005 |
| io.bytes.per.checksum | 512 |
| fs.har.impl.disable.cache | true |
| ipc.client.tcpnodelay | false |
| fs.hftp.impl | org.apache.hadoop.hdfs.HftpFileSystem |
| hadoop.relaxed.worker.version.check | false |
| fs.s3.sleepTimeSeconds | 10 |
| hadoop.http.authentication.type | simple |
No comments:
Post a Comment