blueprint.json 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179
  1. {
  2. "configurations" : [
  3. {
  4. "capacity-scheduler" : {
  5. "yarn.scheduler.capacity.root.default.maximum-am-resource-percent" : "0.5",
  6. "yarn.scheduler.capacity.maximum-am-resource-percent" : "0.5"
  7. }
  8. },
  9. {
  10. "cluster-env": {
  11. "cluster_name": "PigView",
  12. "smokeuser": "ambari-qa",
  13. "user_group": "hadoop",
  14. "security_enabled": "false"
  15. }
  16. },
  17. {
  18. "core-site" : {
  19. "hadoop.proxyuser.root.hosts" : "*",
  20. "hadoop.proxyuser.root.groups" : "*",
  21. "hadoop.proxyuser.hcat.hosts" : "*",
  22. "hadoop.proxyuser.hcat.groups" : "*"
  23. }
  24. },
  25. {
  26. "hadoop-env": {
  27. "dtnode_heapsize" : "250",
  28. "hadoop_heapsize" : "250",
  29. "namenode_heapsize" : "250",
  30. "namenode_opt_newsize": "50",
  31. "namenode_opt_maxnewsize": "100",
  32. "content" : "\n# Set Hadoop-specific environment variables here.\n\n# The only required environment variable is JAVA_HOME. All others are\n# optional. When running a distributed configuration it is best to\n# set JAVA_HOME in this file, so that it is correctly defined on\n# remote nodes.\n\n# The java implementation to use. Required.\nexport JAVA_HOME={{java_home}}\nexport HADOOP_HOME_WARN_SUPPRESS=1\n\n# Hadoop home directory\nexport HADOOP_HOME=${HADOOP_HOME:-/usr/lib/hadoop}\n\n# Hadoop Configuration Directory\n#TODO: if env var set that can cause problems\nexport HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-{{hadoop_conf_dir}}}\n\n{# this is different for HDP1 #}\n# Path to jsvc required by secure HDP 2.0 datanode\nexport JSVC_HOME={{jsvc_path}}\n\n\n# The maximum amount of heap to use, in MB. Default is 1000.\nexport HADOOP_HEAPSIZE=\"{{hadoop_heapsize}}\"\n\nexport HADOOP_NAMENODE_INIT_HEAPSIZE=\"-Xms{{namenode_heapsize}}\"\n\n# Extra Java runtime options. Empty by default.\nexport HADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true ${HADOOP_OPTS}\"\n\n# Command specific options appended to HADOOP_OPTS when specified\nexport HADOOP_NAMENODE_OPTS=\"-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{namenode_opt_newsize}} -XX:MaxNewSize={{namenode_opt_maxnewsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms{{namenode_heapsize}} -Xmx{{namenode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_NAMENODE_OPTS}\"\nHADOOP_JOBTRACKER_OPTS=\"-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{jtnode_opt_newsize}} -XX:MaxNewSize={{jtnode_opt_maxnewsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xmx{{jtnode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dmapred.audit.logger=INFO,MRAUDIT -Dhadoop.mapreduce.jobsummary.logger=INFO,JSA ${HADOOP_JOBTRACKER_OPTS}\"\n\nHADOOP_TASKTRACKER_OPTS=\"-server -Xmx{{ttnode_heapsize}} -Dhadoop.security.logger=ERROR,console -Dmapred.audit.logger=ERROR,console ${HADOOP_TASKTRACKER_OPTS}\"\nHADOOP_DATANODE_OPTS=\"-Xmx{{dtnode_heapsize}} -Dhadoop.security.logger=ERROR,DRFAS ${HADOOP_DATANODE_OPTS}\"\nHADOOP_BALANCER_OPTS=\"-server -Xmx{{hadoop_heapsize}}m ${HADOOP_BALANCER_OPTS}\"\n\nexport HADOOP_SECONDARYNAMENODE_OPTS=\"-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{namenode_opt_newsize}} -XX:MaxNewSize={{namenode_opt_maxnewsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps ${HADOOP_NAMENODE_INIT_HEAPSIZE} -Xmx{{namenode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_SECONDARYNAMENODE_OPTS}\"\n\n# The following applies to multiple commands (fs, dfs, fsck, distcp etc)\nexport HADOOP_CLIENT_OPTS=\"-Xmx${HADOOP_HEAPSIZE}m $HADOOP_CLIENT_OPTS\"\n\n# Extra ssh options. Empty by default.\nexport HADOOP_SSH_OPTS=\"-o ConnectTimeout=5 -o SendEnv=HADOOP_CONF_DIR\"\n\n# Where log files are stored. $HADOOP_HOME/logs by default.\nexport HADOOP_LOG_DIR={{hdfs_log_dir_prefix}}/$USER\n\n# History server logs\nexport HADOOP_MAPRED_LOG_DIR={{mapred_log_dir_prefix}}/$USER\n\n# File naming remote slave hosts. $HADOOP_HOME/conf/slaves by default.\n# export HADOOP_SLAVES=${HADOOP_HOME}/conf/slaves\n\n# host:path where hadoop code should be rsync'd from. Unset by default.\n# export HADOOP_MASTER=master:/home/$USER/src/hadoop\n\n# Seconds to sleep between slave commands. Unset by default. This\n# can be useful in large clusters, where, e.g., slave rsyncs can\n# otherwise arrive faster than the master can service them.\n# export HADOOP_SLAVE_SLEEP=0.1\n\n# The directory where pid files are stored. /tmp by default.\nexport HADOOP_PID_DIR={{hadoop_pid_dir_prefix}}/$USER\n\n# History server pid\nexport HADOOP_MAPRED_PID_DIR={{mapred_pid_dir_prefix}}/$USER\n\nYARN_RESOURCEMANAGER_OPTS=\"-Dyarn.server.resourcemanager.appsummary.logger=INFO,RMSUMMARY\"\n\n# A string representing this instance of hadoop. $USER by default.\nexport HADOOP_IDENT_STRING=$USER\n\n# The scheduling priority for daemon processes. See 'man nice'.\n\n# export HADOOP_NICENESS=10\n\n# Use libraries from standard classpath\nJAVA_JDBC_LIBS=\"\"\n#Add libraries required by mysql connector\nfor jarFile in `ls /usr/share/java/*mysql* 2>/dev/null`\ndo\n JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile\ndone\n#Add libraries required by oracle connector\nfor jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null`\ndo\n JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile\ndone\n#Add libraries required by nodemanager\nMAPREDUCE_LIBS={{mapreduce_libs_path}}\nexport HADOOP_CLASSPATH=${HADOOP_CLASSPATH}${JAVA_JDBC_LIBS}:${MAPREDUCE_LIBS}\n\nif [ -d \"/usr/hdp/current/tez-client\" ]; then\n export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:/usr/hdp/current/tez-client/*:/usr/hdp/current/tez-client/lib/*:/etc/tez/conf\nfi\n\n# Setting path to hdfs command line\nexport HADOOP_LIBEXEC_DIR={{hadoop_libexec_dir}}\n\n#Mostly required for hadoop 2.0\nexport JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}:/usr/lib/hadoop/lib/native/Linux-amd64-64\n\n#Hadoop logging options\nexport HADOOP_ROOT_LOGGER={{hadoop_root_logger}}\n "
  33. }
  34. },
  35. {
  36. "hive-env": {
  37. "content" : "\n if [ \"$SERVICE\" = \"cli\" ]; then\n if [ -z \"$DEBUG\" ]; then\n export HADOOP_OPTS=\"$HADOOP_OPTS -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:+UseParNewGC -XX:-UseGCOverheadLimit\"\n else\n export HADOOP_OPTS=\"$HADOOP_OPTS -XX:NewRatio=12 -Xms10m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:-UseGCOverheadLimit\"\n fi\n fi\n\n# The heap size of the jvm stared by hive shell script can be controlled via:\n\nexport HADOOP_HEAPSIZE=\"{{hive_heapsize}}\"\nexport HADOOP_CLIENT_OPTS=\"-Xmx${HADOOP_HEAPSIZE}m $HADOOP_CLIENT_OPTS\"\n\n# Larger heap size may be required when running queries over large number of files or partitions.\n# By default hive shell scripts use a heap size of 256 (MB). Larger heap size would also be\n# appropriate for hive server (hwi etc).\n\n\n# Set HADOOP_HOME to point to a specific hadoop install directory\nHADOOP_HOME=${HADOOP_HOME:-{{hadoop_home}}}\n\n# Hive Configuration Directory can be controlled by:\nexport HIVE_CONF_DIR={{hive_config_dir}}\n\n# Folder containing extra ibraries required for hive compilation/execution can be controlled by:\nif [ \"${HIVE_AUX_JARS_PATH}\" != \"\" ]; then\n export HIVE_AUX_JARS_PATH=${HIVE_AUX_JARS_PATH}\nelif [ -d \"/usr/lib/hive-webhcat/\" ]; then\n export HIVE_AUX_JARS_PATH=/usr/lib/hive-webhcat/share/hcatalog/hive-hcatalog-core-*.jar\nelse\n export HIVE_AUX_JARS_PATH=/usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core.jar\nfi\nexport METASTORE_PORT={{hive_metastore_port}}\n "
  38. }
  39. },
  40. {
  41. "hbase-site" : {
  42. "hbase.security.authorization": "true",
  43. "hbase.rpc.engine": "org.apache.hadoop.hbase.ipc.SecureRpcEngine",
  44. "hbase_master_heapsize": "250",
  45. "hbase_regionserver_heapsize": "250",
  46. "hbase.rpc.protection": "PRIVACY"
  47. }
  48. },
  49. {
  50. "hdfs-site" : {
  51. "dfs.block.size" : "34217472",
  52. "dfs.replication" : "1",
  53. "dfs.namenode.accesstime.precision" : "3600000",
  54. "dfs.nfs3.dump.dir" : "/tmp/.hdfs-nfs",
  55. "dfs.nfs.exports.allowed.hosts" : "* rw",
  56. "dfs.datanode.max.xcievers" : "1024",
  57. "dfs.block.access.token.enable" : "false"
  58. }
  59. },
  60. {
  61. "hive-site" : {
  62. "javax.jdo.option.ConnectionPassword" : "hive",
  63. "hive.tez.container.size" : "250",
  64. "hive.tez.java.opts" : "-server -Xmx200m -Djava.net.preferIPv4Stack=true",
  65. "hive.heapsize" : "250",
  66. "hive.users.in.admin.role" : "hive",
  67. "hive_metastore_user_passwd" : "hive",
  68. "hive.server2.enable.impersonation": "true",
  69. "hive.compactor.check.interval": "300s",
  70. "hive.compactor.initiator.on": "true",
  71. "hive.compactor.worker.timeout": "86400s",
  72. "hive.enforce.bucketing": "true",
  73. "hive.support.concurrency": "true",
  74. "hive.exec.dynamic.partition.mode": "nonstrict",
  75. "hive.server2.enable.doAs": "false",
  76. "hive.txn.manager": "org.apache.hadoop.hive.ql.lockmgr.DbTxnManager",
  77. "hive.txn.max.open.batch": "1000",
  78. "hive.txn.timeout": "300",
  79. "hive.security.authorization.enabled": "false"
  80. }
  81. },
  82. {
  83. "mapred-env": {
  84. "jobhistory_heapsize" : "250"
  85. }
  86. },
  87. {
  88. "mapred-site" : {
  89. "mapreduce.map.memory.mb" : "250",
  90. "mapreduce.reduce.memory.mb" : "250",
  91. "mapreduce.task.io.sort.mb" : "64",
  92. "yarn.app.mapreduce.am.resource.mb" : "250",
  93. "yarn.app.mapreduce.am.command-opts" : "-Xmx200m",
  94. "mapred.job.reduce.memory.mb" : "250",
  95. "mapred.child.java.opts" : "-Xmx200m",
  96. "mapred.job.map.memory.mb" : "250",
  97. "io.sort.mb" : "64",
  98. "mapreduce.map.java.opts" : "-Xmx200m",
  99. "mapreduce.reduce.java.opts" : "-Xmx200m"
  100. }
  101. },
  102. {
  103. "nagios-env":{
  104. "nagios_contact" : "admin@localhost.localdomain",
  105. "nagios_web_login" : "nagiosadmin",
  106. "nagios_web_password" : "admin"
  107. }
  108. },
  109. {
  110. "tez-site" : {
  111. "tez.am.java.opts" : "-server -Xmx200m -Djava.net.preferIPv4Stack=true -XX:+UseNUMA -XX:+UseParallelGC",
  112. "tez.am.resource.memory.mb" : "250",
  113. "tez.dag.am.resource.memory.mb" : "250",
  114. "yarn.app.mapreduce.am.command-opts" : "-Xmx200m"
  115. }
  116. },
  117. {
  118. "webhcat-site" : {
  119. "webhcat.proxyuser.hcat.hosts" : "*",
  120. "webhcat.proxyuser.hcat.groups" : "*"
  121. }
  122. },
  123. {
  124. "yarn-env": {
  125. "apptimelineserver_heapsize" : "250",
  126. "resourcemanager_heapsize" : "250",
  127. "nodemanager_heapsize" : "250",
  128. "yarn_heapsize" : "250"
  129. }
  130. },
  131. {
  132. "yarn-site" : {
  133. "yarn.nodemanager.resource.memory-mb": "2250",
  134. "yarn.nodemanager.vmem-pmem-ratio" : "10",
  135. "yarn.scheduler.minimum-allocation-mb" : "250",
  136. "yarn.scheduler.maximum-allocation-mb" : "2250",
  137. "yarn.nodemanager.pmem-check-enabled" : "false",
  138. "yarn.acl.enable" : "false",
  139. "yarn.resourcemanager.webapp.proxyuser.hcat.groups" : "*",
  140. "yarn.resourcemanager.webapp.proxyuser.hcat.hosts" : "*"
  141. }
  142. }
  143. ],
  144. "host_groups" : [
  145. {
  146. "name" : "single_host",
  147. "components" : [
  148. {"name": "APP_TIMELINE_SERVER"},
  149. {"name" : "DATANODE"},
  150. {"name" : "GANGLIA_SERVER"},
  151. {"name" : "HDFS_CLIENT"},
  152. {"name" : "HISTORYSERVER"},
  153. {"name" : "HIVE_CLIENT"},
  154. {"name" : "HIVE_METASTORE"},
  155. {"name" : "HIVE_SERVER"},
  156. {"name" : "MAPREDUCE2_CLIENT"},
  157. {"name" : "MYSQL_SERVER"},
  158. {"name" : "NAGIOS_SERVER"},
  159. {"name" : "NAMENODE"},
  160. {"name" : "NODEMANAGER"},
  161. {"name" : "PIG"},
  162. {"name" : "RESOURCEMANAGER"},
  163. {"name" : "SECONDARY_NAMENODE"},
  164. {"name" : "TEZ_CLIENT"},
  165. {"name" : "WEBHCAT_SERVER"},
  166. {"name" : "YARN_CLIENT"},
  167. {"name" : "ZOOKEEPER_CLIENT"},
  168. {"name" : "ZOOKEEPER_SERVER"}
  169. ],
  170. "cardinality" : "1"
  171. }
  172. ],
  173. "Blueprints" : {
  174. "blueprint_name" : "pig-view",
  175. "stack_name" : "HDP",
  176. "stack_version" : "2.2"
  177. }
  178. }