Jelajahi Sumber

AMBARI-1726. It seems upgrades available at the FE is hard-coded to 1.3.0. (yusaku via srimanth)

git-svn-id: https://svn.apache.org/repos/asf/incubator/ambari/trunk@1466296 13f79535-47bb-0310-9956-ffa450edef68
Srimanth 12 tahun lalu
induk
melakukan
8b674ee07a

+ 10 - 7
CHANGES.txt

@@ -14,12 +14,12 @@ Trunk (unreleased changes):
 
 
  AMBARI-1847. Make single PUT call for multiple host overrides. (srimanth)
  AMBARI-1847. Make single PUT call for multiple host overrides. (srimanth)
 
 
- AMBARI-1857. Capacity Scheduler: field order for Add/Edit popup. (srimanth)
+ AMBARI-1857. Capacity Scheduler: field order for Add/Edit popup. (yusaku via srimanth)
 
 
  AMBARI-1855. Capacity Scheduler: when adding a new queue, populate 
  AMBARI-1855. Capacity Scheduler: when adding a new queue, populate 
- fields. (srimanth)
+ fields. (yusaku via srimanth)
  
  
- AMBARI-1850. Update unit tests. (srimanth)
+ AMBARI-1850. Update unit tests. (yusaku via srimanth)
 
 
  AMBARI-1829. HDFS Mirroring: Display Status and handle maintenance operations 
  AMBARI-1829. HDFS Mirroring: Display Status and handle maintenance operations 
  like Stop, Suspend, Activate etc. (Arun Kandregula via srimanth)
  like Stop, Suspend, Activate etc. (Arun Kandregula via srimanth)
@@ -677,15 +677,18 @@ Trunk (unreleased changes):
 
 
  BUG FIXES
  BUG FIXES
 
 
- AMBARI-1854. Wizards available for a non-administrator user. (srimanth)
+ AMBARI-1726. It seems upgrades available at the FE is hard-coded to 1.3.0. 
+ (yusaku via srimanth)
+
+ AMBARI-1854. Wizards available for a non-administrator user. (yusaku via srimanth)
 
 
  AMBARI-1852. Upon clicking Services > Service > Config, a call to 
  AMBARI-1852. Upon clicking Services > Service > Config, a call to 
- "configurations resource is made and the server throws 400. (srimanth)
+ "configurations resource is made and the server throws 400. (yusaku via srimanth)
 
 
  AMBARI-1851. Ambari Web behaves strangely when there is no Active HBase 
  AMBARI-1851. Ambari Web behaves strangely when there is no Active HBase 
- Master. (srimanth)
+ Master. (yusaku via srimanth)
 
 
- AMBARI-1849. Cosmetic problems on HBase Dashboard. (srimanth)
+ AMBARI-1849. Cosmetic problems on HBase Dashboard. (yusaku via srimanth)
 
 
  AMBARI-1848. Install Wizard, Step 7: Oozie Database Derby option should say 
  AMBARI-1848. Install Wizard, Step 7: Oozie Database Derby option should say 
  "New Derby Database", not "Current Derby Database". (Xi Wang via srimanth)
  "New Derby Database", not "Current Derby Database". (Xi Wang via srimanth)

+ 1 - 1
ambari-web/app/app.js

@@ -42,7 +42,7 @@ module.exports = Em.Application.create({
    */
    */
   stack2VersionURL:function(){
   stack2VersionURL:function(){
     var stackVersion = this.get('currentStackVersion') || this.get('defaultStackVersion');
     var stackVersion = this.get('currentStackVersion') || this.get('defaultStackVersion');
-    return '/stacks2/HDP/versions/1.3.0';//TODO + stackVersion.replace(/HDP-/g, '');
+    return '/stacks2/HDP/versions/' + stackVersion.replace(/HDP-/g, '');
   }.property('currentStackVersion'),
   }.property('currentStackVersion'),
   clusterName: null,
   clusterName: null,
   currentStackVersion: null
   currentStackVersion: null

+ 4 - 4
ambari-web/app/assets/data/dashboard/services.json

@@ -28,7 +28,7 @@
                 "cluster_name" : "cl1",
                 "cluster_name" : "cl1",
                 "desired_configs" : { },
                 "desired_configs" : { },
                 "desired_state" : "INSTALLED",
                 "desired_state" : "INSTALLED",
-                "state" : "INSTALLED",
+                "state" : "STARTED",
                 "component_name" : "HBASE_REGIONSERVER",
                 "component_name" : "HBASE_REGIONSERVER",
                 "host_name" : "domU-12-31-39-0E-E6-01.compute-1.internal"
                 "host_name" : "domU-12-31-39-0E-E6-01.compute-1.internal"
               }
               }
@@ -43,7 +43,7 @@
                 "cluster_name" : "cl1",
                 "cluster_name" : "cl1",
                 "desired_configs" : { },
                 "desired_configs" : { },
                 "desired_state" : "INSTALLED",
                 "desired_state" : "INSTALLED",
-                "state" : "INSTALLED",
+                "state" : "STARTED",
                 "component_name" : "HBASE_REGIONSERVER",
                 "component_name" : "HBASE_REGIONSERVER",
                 "host_name" : "domU-12-31-39-16-48-4B.compute-1.internal"
                 "host_name" : "domU-12-31-39-16-48-4B.compute-1.internal"
               }
               }
@@ -1024,7 +1024,7 @@
                 "cluster_name" : "tdk",
                 "cluster_name" : "tdk",
                 "ha_status" : "passive",
                 "ha_status" : "passive",
                 "desired_state" : "INSTALLED",
                 "desired_state" : "INSTALLED",
-                "state" : "St",
+                "state" : "STARTED",
                 "component_name" : "HUE_SERVER",
                 "component_name" : "HUE_SERVER",
                 "service_name" : "HUE",
                 "service_name" : "HUE",
                 "host_name" : "ip-10-147-169-187.ec2.internal",
                 "host_name" : "ip-10-147-169-187.ec2.internal",
@@ -1059,7 +1059,7 @@
                 "cluster_name" : "tdk",
                 "cluster_name" : "tdk",
                 "ha_status" : "passive",
                 "ha_status" : "passive",
                 "desired_state" : "INSTALLED",
                 "desired_state" : "INSTALLED",
-                "state" : "St",
+                "state" : "STARTED",
                 "component_name" : "HUE_SERVER",
                 "component_name" : "HUE_SERVER",
                 "service_name" : "HUE",
                 "service_name" : "HUE",
                 "host_name" : "ip-10-147-169-187.ec2.internal",
                 "host_name" : "ip-10-147-169-187.ec2.internal",

+ 3 - 19
ambari-web/app/assets/data/wizard/stack/hdp/version130/HCATALOG.json

@@ -1,20 +1,4 @@
 {
 {
-  "name" : "HCATALOG",
-  "version" : "0.5.0",
-  "user" : "root",
-  "comment" : "This is comment for HCATALOG service",
-  "properties" : [ ],
-  "components" : [ {
-    "name" : "HCAT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : true,
-  "clientComponent" : {
-    "name" : "HCAT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HCATALOG/configurations?fields=*",
+  "items" : [ ]
+}

+ 736 - 209
ambari-web/app/assets/data/wizard/stack/hdp/version130/HDFS.json

@@ -1,210 +1,737 @@
 {
 {
-  "name" : "HDFS",
-  "version" : "1.1.2",
-  "user" : "root",
-  "comment" : "Apache Hadoop Distributed File System",
-  "properties" : [ {
-    "name" : "dfs.datanode.socket.write.timeout",
-    "value" : "0",
-    "description" : "DFS Client write socket timeout",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.replication.max",
-    "value" : "50",
-    "description" : "Maximal block replication.\n  ",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.heartbeat.interval",
-    "value" : "3",
-    "description" : "Determines datanode heartbeat interval in seconds.",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.safemode.threshold.pct",
-    "value" : "1.0f",
-    "description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.balance.bandwidthPerSec",
-    "value" : "6250000",
-    "description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.block.size",
-    "value" : "134217728",
-    "description" : "The default block size for new files.",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.datanode.ipc.address",
-    "value" : "0.0.0.0:8010",
-    "description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.blockreport.initialDelay",
-    "value" : "120",
-    "description" : "Delay for first block report in seconds.",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.datanode.du.pct",
-    "value" : "0.85f",
-    "description" : "When calculating remaining space, only use this percentage of the real available space\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.namenode.handler.count",
-    "value" : "40",
-    "description" : "The number of server threads for the namenode.",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.datanode.max.xcievers",
-    "value" : "4096",
-    "description" : "PRIVATE CONFIG VARIABLE",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.umaskmode",
-    "value" : "077",
-    "description" : "\nThe octal umask used when creating files and directories.\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.web.ugi",
-    "value" : "gopher,gopher",
-    "description" : "The user account used by the web interface.\nSyntax: USERNAME,GROUP1,GROUP2, ...\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.permissions",
-    "value" : "true",
-    "description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.permissions.supergroup",
-    "value" : "hdfs",
-    "description" : "The name of the group of super-users.",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.namenode.handler.count",
-    "value" : "100",
-    "description" : "Added to grow Queue size so that more client connections are allowed",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "ipc.server.max.response.size",
-    "value" : "5242880",
-    "description" : null,
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.block.access.token.enable",
-    "value" : "true",
-    "description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.secondary.https.port",
-    "value" : "50490",
-    "description" : "The https port where secondary-namenode binds",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.https.port",
-    "value" : "50470",
-    "description" : "The https port where namenode binds",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.access.time.precision",
-    "value" : "0",
-    "description" : "The access time for HDFS file is precise upto this value.\n               The default value is 1 hour. Setting a value of 0 disables\n               access times for HDFS.\n  ",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.cluster.administrators",
-    "value" : " hdfs",
-    "description" : "ACL for who all can view the default servlets in the HDFS",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "ipc.server.read.threadpool.size",
-    "value" : "5",
-    "description" : null,
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "dfs.datanode.failed.volumes.tolerated",
-    "value" : "0",
-    "description" : "Number of failed disks datanode would tolerate",
-    "filename" : "hdfs-site.xml"
-  }, {
-    "name" : "io.file.buffer.size",
-    "value" : "131072",
-    "description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "io.serializations",
-    "value" : "org.apache.hadoop.io.serializer.WritableSerialization",
-    "description" : null,
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "io.compression.codec.lzo.class",
-    "value" : "com.hadoop.compression.lzo.LzoCodec",
-    "description" : "The implementation for lzo codec.",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "fs.trash.interval",
-    "value" : "360",
-    "description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "fs.checkpoint.edits.dir",
-    "value" : "${fs.checkpoint.dir}",
-    "description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "fs.checkpoint.period",
-    "value" : "21600",
-    "description" : "The number of seconds between two periodic checkpoints.\n  ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "fs.checkpoint.size",
-    "value" : "536870912",
-    "description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "ipc.client.idlethreshold",
-    "value" : "8000",
-    "description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "ipc.client.connection.maxidletime",
-    "value" : "30000",
-    "description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "ipc.client.connect.max.retries",
-    "value" : "50",
-    "description" : "Defines the maximum number of retries for IPC connections.",
-    "filename" : "core-site.xml"
-  }, {
-    "name" : "webinterface.private.actions",
-    "value" : "false",
-    "description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
-    "filename" : "core-site.xml"
-  } ],
-  "components" : [ {
-    "name" : "NAMENODE",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "DATANODE",
-    "category" : "SLAVE",
-    "client" : false,
-    "master" : false
-  }, {
-    "name" : "SECONDARY_NAMENODE",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "HDFS_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "HDFS_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.blockreport.initialDelay",
+      "StackConfigurations" : {
+        "property_description" : "Delay for first block report in seconds.",
+        "property_value" : "120",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.blockreport.initialDelay",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.safemode.threshold.pct",
+      "StackConfigurations" : {
+        "property_description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
+        "property_value" : "1.0f",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.safemode.threshold.pct",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_dir",
+      "StackConfigurations" : {
+        "property_description" : "Secondary NameNode checkpoint dir.",
+        "property_value" : "/hadoop/hdfs/namesecondary",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs_checkpoint_dir",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.umaskmode",
+      "StackConfigurations" : {
+        "property_description" : "\nThe octal umask used when creating files and directories.\n",
+        "property_value" : "077",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.umaskmode",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.compression.codec.lzo.class",
+      "StackConfigurations" : {
+        "property_description" : "The implementation for lzo codec.",
+        "property_value" : "com.hadoop.compression.lzo.LzoCodec",
+        "stack_version" : "1.3.0",
+        "property_name" : "io.compression.codec.lzo.class",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.heartbeat.interval",
+      "StackConfigurations" : {
+        "property_description" : "Determines datanode heartbeat interval in seconds.",
+        "property_value" : "3",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.heartbeat.interval",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_datanode_failed_volume_tolerated",
+      "StackConfigurations" : {
+        "property_description" : "DataNode volumes failure toleration",
+        "property_value" : "0",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs_datanode_failed_volume_tolerated",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_name_dir",
+      "StackConfigurations" : {
+        "property_description" : "NameNode Directories.",
+        "property_value" : "/hadoop/hdfs/namenode",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs_name_dir",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_size",
+      "StackConfigurations" : {
+        "property_description" : "FS Checkpoint Size.",
+        "property_value" : "0.5",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs_checkpoint_size",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.balance.bandwidthPerSec",
+      "StackConfigurations" : {
+        "property_description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
+        "property_value" : "6250000",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.balance.bandwidthPerSec",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_maxnewsize",
+      "StackConfigurations" : {
+        "property_description" : "NameNode maximum new generation size",
+        "property_value" : "640",
+        "stack_version" : "1.3.0",
+        "property_name" : "namenode_opt_maxnewsize",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.edits.dir",
+      "StackConfigurations" : {
+        "property_description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
+        "property_value" : "${fs.checkpoint.dir}",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs.checkpoint.edits.dir",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/keytab_path",
+      "StackConfigurations" : {
+        "property_description" : "KeyTab Directory.",
+        "property_value" : "/etc/security/keytabs",
+        "stack_version" : "1.3.0",
+        "property_name" : "keytab_path",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.size",
+      "StackConfigurations" : {
+        "property_description" : "The default block size for new files.",
+        "property_value" : "134217728",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.block.size",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security_enabled",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop Security",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "security_enabled",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.serializations",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "org.apache.hadoop.io.serializer.WritableSerialization",
+        "stack_version" : "1.3.0",
+        "property_name" : "io.serializations",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.task.umbilical.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for TaskUmbilicalProtocol, used by the map and reduce\n    tasks to communicate with the parent tasktracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.task.umbilical.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/kerberos_domain",
+      "StackConfigurations" : {
+        "property_description" : "Kerberos realm.",
+        "property_value" : "EXAMPLE.COM",
+        "stack_version" : "1.3.0",
+        "property_name" : "kerberos_domain",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_data_dir",
+      "StackConfigurations" : {
+        "property_description" : "Data directories for Data Nodes.",
+        "property_value" : "/hadoop/hdfs/data",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs_data_dir",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hdfs_log_dir_prefix",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop Log Dir Prefix",
+        "property_value" : "/var/log/hadoop",
+        "stack_version" : "1.3.0",
+        "property_name" : "hdfs_log_dir_prefix",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.datanode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for DatanodeProtocol, which is used by datanodes to\n    communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.datanode.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connect.max.retries",
+      "StackConfigurations" : {
+        "property_description" : "Defines the maximum number of retries for IPC connections.",
+        "property_value" : "50",
+        "stack_version" : "1.3.0",
+        "property_name" : "ipc.client.connect.max.retries",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs_checkpoint_period",
+      "StackConfigurations" : {
+        "property_description" : "HDFS Maximum Checkpoint Delay",
+        "property_value" : "21600",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs_checkpoint_period",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "The number of server threads for the namenode.",
+        "property_value" : "40",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.namenode.handler.count",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.ipc.address",
+      "StackConfigurations" : {
+        "property_description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
+        "property_value" : "0.0.0.0:8010",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.datanode.ipc.address",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs_webhdfs_enabled",
+      "StackConfigurations" : {
+        "property_description" : "WebHDFS enabled",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs_webhdfs_enabled",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.trash.interval",
+      "StackConfigurations" : {
+        "property_description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
+        "property_value" : "360",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs.trash.interval",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.secondary.https.port",
+      "StackConfigurations" : {
+        "property_description" : "The https port where secondary-namenode binds",
+        "property_value" : "50490",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.secondary.https.port",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/datanode_du_reserved",
+      "StackConfigurations" : {
+        "property_description" : "Reserved space for HDFS",
+        "property_value" : "1",
+        "stack_version" : "1.3.0",
+        "property_name" : "datanode_du_reserved",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/io.file.buffer.size",
+      "StackConfigurations" : {
+        "property_description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
+        "property_value" : "131072",
+        "stack_version" : "1.3.0",
+        "property_name" : "io.file.buffer.size",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.replication.max",
+      "StackConfigurations" : {
+        "property_description" : "Maximal block replication.\n  ",
+        "property_value" : "50",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.replication.max",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_pid_dir_prefix",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop PID Dir Prefix",
+        "property_value" : "/var/run/hadoop",
+        "stack_version" : "1.3.0",
+        "property_name" : "hadoop_pid_dir_prefix",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.datanode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for InterDatanodeProtocol, the inter-datanode protocol\n    for updating generation timestamp.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.inter.datanode.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.socket.write.timeout",
+      "StackConfigurations" : {
+        "property_description" : "DFS Client write socket timeout",
+        "property_value" : "0",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.datanode.socket.write.timeout",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.max.xcievers",
+      "StackConfigurations" : {
+        "property_description" : "PRIVATE CONFIG VARIABLE",
+        "property_value" : "4096",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.datanode.max.xcievers",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.max.response.size",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "5242880",
+        "stack_version" : "1.3.0",
+        "property_name" : "ipc.server.max.response.size",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.size",
+      "StackConfigurations" : {
+        "property_description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
+        "property_value" : "536870912",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs.checkpoint.size",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.namenode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for NamenodeProtocol, the protocol used by the secondary\n    namenode to communicate with the namenode.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.namenode.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions",
+      "StackConfigurations" : {
+        "property_description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.permissions",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.https.port",
+      "StackConfigurations" : {
+        "property_description" : "The https port where namenode binds",
+        "property_value" : "50470",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.https.port",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_heapsize",
+      "StackConfigurations" : {
+        "property_description" : "NameNode Java heap size",
+        "property_value" : "1024",
+        "stack_version" : "1.3.0",
+        "property_name" : "namenode_heapsize",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.namenode.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "Added to grow Queue size so that more client connections are allowed",
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.namenode.handler.count",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.web.ugi",
+      "StackConfigurations" : {
+        "property_description" : "The user account used by the web interface.\nSyntax: USERNAME,GROUP1,GROUP2, ...\n",
+        "property_value" : "gopher,gopher",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.web.ugi",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.du.pct",
+      "StackConfigurations" : {
+        "property_description" : "When calculating remaining space, only use this percentage of the real available space\n",
+        "property_value" : "0.85f",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.datanode.du.pct",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/fs.checkpoint.period",
+      "StackConfigurations" : {
+        "property_description" : "The number of seconds between two periodic checkpoints.\n  ",
+        "property_value" : "21600",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs.checkpoint.period",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.block.access.token.enable",
+      "StackConfigurations" : {
+        "property_description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.block.access.token.enable",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.cluster.administrators",
+      "StackConfigurations" : {
+        "property_description" : "ACL for who all can view the default servlets in the HDFS",
+        "property_value" : " hdfs",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.cluster.administrators",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dtnode_heapsize",
+      "StackConfigurations" : {
+        "property_description" : "DataNode maximum Java heap size",
+        "property_value" : "1024",
+        "stack_version" : "1.3.0",
+        "property_name" : "dtnode_heapsize",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.job.submission.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for JobSubmissionProtocol, used by job clients to\n    communciate with the jobtracker for job submission, querying job status etc.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.job.submission.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/webinterface.private.actions",
+      "StackConfigurations" : {
+        "property_description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "webinterface.private.actions",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.connection.maxidletime",
+      "StackConfigurations" : {
+        "property_description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
+        "property_value" : "30000",
+        "stack_version" : "1.3.0",
+        "property_name" : "ipc.client.connection.maxidletime",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.permissions.supergroup",
+      "StackConfigurations" : {
+        "property_description" : "The name of the group of super-users.",
+        "property_value" : "hdfs",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.permissions.supergroup",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/hadoop_heapsize",
+      "StackConfigurations" : {
+        "property_description" : "Hadoop maximum Java heap size",
+        "property_value" : "1024",
+        "stack_version" : "1.3.0",
+        "property_name" : "hadoop_heapsize",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.client.idlethreshold",
+      "StackConfigurations" : {
+        "property_description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
+        "property_value" : "8000",
+        "stack_version" : "1.3.0",
+        "property_name" : "ipc.client.idlethreshold",
+        "filename" : "core-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.inter.tracker.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for InterTrackerProtocol, used by the tasktrackers to\n    communicate with the jobtracker.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.inter.tracker.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.datanode.failed.volumes.tolerated",
+      "StackConfigurations" : {
+        "property_description" : "Number of failed disks datanode would tolerate",
+        "property_value" : "0",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.datanode.failed.volumes.tolerated",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/namenode_opt_newsize",
+      "StackConfigurations" : {
+        "property_description" : "NameNode new generation size",
+        "property_value" : "200",
+        "stack_version" : "1.3.0",
+        "property_name" : "namenode_opt_newsize",
+        "filename" : "global.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.datanode.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for ClientDatanodeProtocol, the client-to-datanode protocol\n    for block recovery.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.client.datanode.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/security.client.protocol.acl",
+      "StackConfigurations" : {
+        "property_description" : "ACL for ClientProtocol, which is used by user code\n    via the DistributedFileSystem.\n    The ACL is a comma-separated list of user and group names. The user and\n    group list is separated by a blank. For e.g. \"alice,bob users,wheel\".\n    A special value of \"*\" means all users are allowed.",
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "security.client.protocol.acl",
+        "filename" : "hadoop-policy.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/ipc.server.read.threadpool.size",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "5",
+        "stack_version" : "1.3.0",
+        "property_name" : "ipc.server.read.threadpool.size",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS/configurations/dfs.access.time.precision",
+      "StackConfigurations" : {
+        "property_description" : "The access time for HDFS file is precise upto this value.\n               The default value is 1 hour. Setting a value of 0 disables\n               access times for HDFS.\n  ",
+        "property_value" : "0",
+        "stack_version" : "1.3.0",
+        "property_name" : "dfs.access.time.precision",
+        "filename" : "hdfs-site.xml",
+        "service_name" : "HDFS",
+        "stack_name" : "HDP"
+      }
+    }
+  ]
+}

+ 208 - 94
ambari-web/app/assets/data/wizard/stack/hdp/version130/HIVE.json

@@ -1,95 +1,209 @@
 {
 {
-  "name" : "HIVE",
-  "version" : "0.10.0",
-  "user" : "root",
-  "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-  "properties" : [ {
-    "name" : "hive.metastore.local",
-    "value" : "false",
-    "description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "javax.jdo.option.ConnectionDriverName",
-    "value" : "com.mysql.jdbc.Driver",
-    "description" : "Driver class name for a JDBC metastore",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.metastore.warehouse.dir",
-    "value" : "/apps/hive/warehouse",
-    "description" : "location of default database for the warehouse",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.metastore.cache.pinobjtypes",
-    "value" : "Table,Database,Type,FieldSchema,Order",
-    "description" : "List of comma separated metastore object types that should be pinned in the cache",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.semantic.analyzer.factory.impl",
-    "value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
-    "description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hadoop.clientside.fs.operations",
-    "value" : "true",
-    "description" : "FS operations are owned by client",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.metastore.client.socket.timeout",
-    "value" : "60",
-    "description" : "MetaStore Client socket timeout in seconds",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.metastore.execute.setugi",
-    "value" : "true",
-    "description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.security.authorization.enabled",
-    "value" : "true",
-    "description" : "enable or disable the hive client authorization",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.security.authorization.manager",
-    "value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
-    "description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "hive.server2.enable.doAs",
-    "value" : "true",
-    "description" : null,
-    "filename" : "hive-site.xml"
-  }, {
-    "name" : "fs.hdfs.impl.disable.cache",
-    "value" : "true",
-    "description" : null,
-    "filename" : "hive-site.xml"
-  } ],
-  "components" : [ {
-    "name" : "HIVE_METASTORE",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "HIVE_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "MYSQL_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "HIVE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "HIVE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/javax.jdo.option.ConnectionDriverName",
+      "StackConfigurations" : {
+        "property_description" : "Driver class name for a JDBC metastore",
+        "property_value" : "com.mysql.jdbc.Driver",
+        "stack_version" : "1.3.0",
+        "property_name" : "javax.jdo.option.ConnectionDriverName",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_pid_dir",
+      "StackConfigurations" : {
+        "property_description" : "Hive PID Dir.",
+        "property_value" : "/var/run/hive",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive_pid_dir",
+        "filename" : "global.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.enabled",
+      "StackConfigurations" : {
+        "property_description" : "enable or disable the hive client authorization",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.security.authorization.enabled",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_conf_dir",
+      "StackConfigurations" : {
+        "property_description" : "Hive Conf Dir.",
+        "property_value" : "/etc/hive/conf",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive_conf_dir",
+        "filename" : "global.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hadoop.clientside.fs.operations",
+      "StackConfigurations" : {
+        "property_description" : "FS operations are owned by client",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "hadoop.clientside.fs.operations",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.security.authorization.manager",
+      "StackConfigurations" : {
+        "property_description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
+        "property_value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.security.authorization.manager",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/fs.hdfs.impl.disable.cache",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "fs.hdfs.impl.disable.cache",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.warehouse.dir",
+      "StackConfigurations" : {
+        "property_description" : "location of default database for the warehouse",
+        "property_value" : "/apps/hive/warehouse",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.metastore.warehouse.dir",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.semantic.analyzer.factory.impl",
+      "StackConfigurations" : {
+        "property_description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
+        "property_value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.semantic.analyzer.factory.impl",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_aux_jars_path",
+      "StackConfigurations" : {
+        "property_description" : "Hive auxiliary jar path.",
+        "property_value" : "/usr/lib/hcatalog/share/hcatalog/hcatalog-core.jar",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive_aux_jars_path",
+        "filename" : "global.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.client.socket.timeout",
+      "StackConfigurations" : {
+        "property_description" : "MetaStore Client socket timeout in seconds",
+        "property_value" : "60",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.metastore.client.socket.timeout",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.server2.enable.doAs",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.server2.enable.doAs",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.cache.pinobjtypes",
+      "StackConfigurations" : {
+        "property_description" : "List of comma separated metastore object types that should be pinned in the cache",
+        "property_value" : "Table,Database,Type,FieldSchema,Order",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.metastore.cache.pinobjtypes",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.execute.setugi",
+      "StackConfigurations" : {
+        "property_description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.metastore.execute.setugi",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/mysql_connector_url",
+      "StackConfigurations" : {
+        "property_description" : "Hive PID Dir.",
+        "property_value" : "${download_url}/mysql-connector-java-5.1.18.zip",
+        "stack_version" : "1.3.0",
+        "property_name" : "mysql_connector_url",
+        "filename" : "global.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive.metastore.local",
+      "StackConfigurations" : {
+        "property_description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive.metastore.local",
+        "filename" : "hive-site.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE/configurations/hive_log_dir",
+      "StackConfigurations" : {
+        "property_description" : "Directory for Hive Log files.",
+        "property_value" : "/var/log/hive",
+        "stack_version" : "1.3.0",
+        "property_name" : "hive_log_dir",
+        "filename" : "global.xml",
+        "service_name" : "HIVE",
+        "stack_name" : "HDP"
+      }
+    }
+  ]
+}

+ 772 - 229
ambari-web/app/assets/data/wizard/stack/hdp/version130/MAPREDUCE.json

@@ -1,230 +1,773 @@
 {
 {
-  "name" : "MAPREDUCE",
-  "version" : "1.1.2",
-  "user" : "mapred",
-  "comment" : "Apache Hadoop Distributed Processing Framework",
-  "properties" : [ {
-    "name" : "io.sort.record.percent",
-    "value" : ".2",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "io.sort.factor",
-    "value" : "100",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
-    "value" : "250",
-    "description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.tracker.handler.count",
-    "value" : "50",
-    "description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.system.dir",
-    "value" : "/mapred/system",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.cluster.administrators",
-    "value" : " hadoop",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.reduce.parallel.copies",
-    "value" : "30",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "tasktracker.http.threads",
-    "value" : "50",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.map.tasks.speculative.execution",
-    "value" : "false",
-    "description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.reduce.tasks.speculative.execution",
-    "value" : "false",
-    "description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.reduce.slowstart.completed.maps",
-    "value" : "0.05",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.inmem.merge.threshold",
-    "value" : "1000",
-    "description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.shuffle.merge.percent",
-    "value" : "0.66",
-    "description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.shuffle.input.buffer.percent",
-    "value" : "0.7",
-    "description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.output.compression.type",
-    "value" : "BLOCK",
-    "description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.completeuserjobs.maximum",
-    "value" : "0",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.restart.recover",
-    "value" : "false",
-    "description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.reduce.input.buffer.percent",
-    "value" : "0.0",
-    "description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.reduce.input.limit",
-    "value" : "10737418240",
-    "description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.task.timeout",
-    "value" : "600000",
-    "description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "jetty.connector",
-    "value" : "org.mortbay.jetty.nio.SelectChannelConnector",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.child.root.logger",
-    "value" : "INFO,TLA",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.max.tracker.blacklists",
-    "value" : "16",
-    "description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.healthChecker.interval",
-    "value" : "135000",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.healthChecker.script.timeout",
-    "value" : "60000",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.tracker.persist.jobstatus.active",
-    "value" : "false",
-    "description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.tracker.persist.jobstatus.hours",
-    "value" : "1",
-    "description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.retirejob.check",
-    "value" : "10000",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.retirejob.interval",
-    "value" : "0",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.tracker.history.completed.location",
-    "value" : "/mapred/history/done",
-    "description" : "No description",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
-    "value" : "false",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.job.reuse.jvm.num.tasks",
-    "value" : "1",
-    "description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "hadoop.job.history.user.location",
-    "value" : "none",
-    "description" : null,
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.jobtracker.staging.root.dir",
-    "value" : "/user",
-    "description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n   name. It is a path in the default file system.",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.tasktracker.group",
-    "value" : "hadoop",
-    "description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.jobtracker.split.metainfo.maxsize",
-    "value" : "50000000",
-    "description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapreduce.history.server.embedded",
-    "value" : "false",
-    "description" : "Should job history server be embedded within Job tracker\nprocess",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.blacklist.fault-timeout-window",
-    "value" : "180",
-    "description" : "\n    3-hour sliding window (value is in minutes)\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.jobtracker.blacklist.fault-bucket-width",
-    "value" : "15",
-    "description" : "\n    15-minute bucket size (value is in minutes)\n  ",
-    "filename" : "mapred-site.xml"
-  }, {
-    "name" : "mapred.queue.names",
-    "value" : "default",
-    "description" : " Comma separated list of queues configured for this jobtracker.",
-    "filename" : "mapred-site.xml"
-  } ],
-  "components" : [ {
-    "name" : "JOBTRACKER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "TASKTRACKER",
-    "category" : "SLAVE",
-    "client" : false,
-    "master" : false
-  }, {
-    "name" : "MAPREDUCE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "MAPREDUCE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.tasks.speculative.execution",
+      "StackConfigurations" : {
+        "property_description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.reduce.tasks.speculative.execution",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
+      "StackConfigurations" : {
+        "property_description" : "The default maximum number of tasks per-user, across all the of \n    the user's jobs in the queue, which can be initialized concurrently. Once \n    the user's jobs exceed this limit they will be queued on disk.  \n    ",
+        "property_value" : "100000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-user",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
+      "StackConfigurations" : {
+        "property_description" : "The multipe of (maximum-system-jobs * queue-capacity) used to \n    determine the number of jobs which are accepted by the scheduler.  \n    ",
+        "property_value" : "10",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.init-accept-jobs-factor",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.init-poll-interval",
+      "StackConfigurations" : {
+        "property_description" : "The amount of time in miliseconds which is used to poll \n    the job queues for jobs to initialize.\n    ",
+        "property_value" : "5000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.init-poll-interval",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.history.server.embedded",
+      "StackConfigurations" : {
+        "property_description" : "Should job history server be embedded within Job tracker\nprocess",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.history.server.embedded",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-bucket-width",
+      "StackConfigurations" : {
+        "property_description" : "\n    15-minute bucket size (value is in minutes)\n  ",
+        "property_value" : "15",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.blacklist.fault-bucket-width",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.map.tasks.speculative.execution",
+      "StackConfigurations" : {
+        "property_description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.map.tasks.speculative.execution",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.supports-priority",
+      "StackConfigurations" : {
+        "property_description" : "If true, priorities of jobs will be taken into \n      account in scheduling decisions.\n    ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.supports-priority",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.blacklist.fault-timeout-window",
+      "StackConfigurations" : {
+        "property_description" : "\n    3-hour sliding window (value is in minutes)\n  ",
+        "property_value" : "180",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.blacklist.fault-timeout-window",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.active",
+      "StackConfigurations" : {
+        "property_description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.tracker.persist.jobstatus.active",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-supports-priority",
+      "StackConfigurations" : {
+        "property_description" : "If true, priorities of jobs will be taken into \n      account in scheduling decisions by default in a job queue.\n    ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-supports-priority",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-user-limit-factor",
+      "StackConfigurations" : {
+        "property_description" : "The default multiple of queue-capacity which is used to \n    determine the amount of slots a single user can consume concurrently.\n    ",
+        "property_value" : "1",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-user-limit-factor",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.record.percent",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : ".2",
+        "stack_version" : "1.3.0",
+        "property_name" : "io.sort.record.percent",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.check",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "10000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.retirejob.check",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.system.dir",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : "/mapred/system",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.system.dir",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-capacity",
+      "StackConfigurations" : {
+        "property_description" : "\n\tmaximum-capacity defines a limit beyond which a queue cannot use the capacity of the cluster.\n\tThis provides a means to limit how much excess capacity a queue can use. By default, there is no limit.\n\tThe maximum-capacity of a queue can only be greater than or equal to its minimum capacity.\n        Default value of -1 implies a queue can use complete capacity of the cluster.\n\n        This property could be to curtail certain jobs which are long running in nature from occupying more than a \n        certain percentage of the cluster, which in the absence of pre-emption, could lead to capacity guarantees of \n        other queues being affected.\n        \n        One important thing to note is that maximum-capacity is a percentage , so based on the cluster's capacity\n        the max capacity would change. So if large no of nodes or racks get added to the cluster , max Capacity in \n        absolute terms would increase accordingly.\n    ",
+        "property_value" : "-1",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-capacity",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jetty.connector",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : "org.mortbay.jetty.nio.SelectChannelConnector",
+        "stack_version" : "1.3.0",
+        "property_name" : "jetty.connector",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
+      "StackConfigurations" : {
+        "property_description" : "The default maximum number of tasks, across all jobs in the \n    queue, which can be initialized concurrently. Once the queue's jobs exceed \n    this limit they will be queued on disk.  \n    ",
+        "property_value" : "200000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-maximum-active-tasks-per-queue",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/scheduler_name",
+      "StackConfigurations" : {
+        "property_description" : "MapRed Capacity Scheduler.",
+        "property_value" : "org.apache.hadoop.mapred.CapacityTaskScheduler",
+        "stack_version" : "1.3.0",
+        "property_name" : "scheduler_name",
+        "filename" : "global.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reuse.jvm.num.tasks",
+      "StackConfigurations" : {
+        "property_description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
+        "property_value" : "1",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.reuse.jvm.num.tasks",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.max.tracker.blacklists",
+      "StackConfigurations" : {
+        "property_description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
+        "property_value" : "16",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.max.tracker.blacklists",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-submit-job",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.queue.default.acl-submit-job",
+        "filename" : "mapred-queue-acls.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.names",
+      "StackConfigurations" : {
+        "property_description" : " Comma separated list of queues configured for this jobtracker.",
+        "property_value" : "default",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.queue.names",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.reduce.input.buffer.percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
+        "property_value" : "0.0",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.reduce.input.buffer.percent",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobtracker.split.metainfo.maxsize",
+      "StackConfigurations" : {
+        "property_description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
+        "property_value" : "50000000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.jobtracker.split.metainfo.maxsize",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/hadoop.job.history.user.location",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "none",
+        "stack_version" : "1.3.0",
+        "property_name" : "hadoop.job.history.user.location",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.completeuserjobs.maximum",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.completeuserjobs.maximum",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.output.compression.type",
+      "StackConfigurations" : {
+        "property_description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
+        "property_value" : "BLOCK",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.output.compression.type",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.retirejob.interval",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "0",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.retirejob.interval",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.interval",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "135000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.healthChecker.interval",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_opt_newsize",
+      "StackConfigurations" : {
+        "property_description" : "MapRed Capacity Scheduler.",
+        "property_value" : "200",
+        "stack_version" : "1.3.0",
+        "property_name" : "jtnode_opt_newsize",
+        "filename" : "global.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.task.timeout",
+      "StackConfigurations" : {
+        "property_description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
+        "property_value" : "600000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.task.timeout",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.persist.jobstatus.hours",
+      "StackConfigurations" : {
+        "property_description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
+        "property_value" : "1",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.tracker.persist.jobstatus.hours",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.merge.percent",
+      "StackConfigurations" : {
+        "property_description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
+        "property_value" : "0.66",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.shuffle.merge.percent",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.jobtracker.restart.recover",
+      "StackConfigurations" : {
+        "property_description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.jobtracker.restart.recover",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.handler.count",
+      "StackConfigurations" : {
+        "property_description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
+        "property_value" : "50",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.tracker.handler.count",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
+      "StackConfigurations" : {
+        "property_description" : "The maximum number of tasks per-user, across all the of the \n    user's jobs in the queue, which can be initialized concurrently. Once the \n    user's jobs exceed this limit they will be queued on disk.  \n    ",
+        "property_value" : "100000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks-per-user",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.tasktracker.group",
+      "StackConfigurations" : {
+        "property_description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
+        "property_value" : "hadoop",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.tasktracker.group",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
+      "StackConfigurations" : {
+        "property_description" : "The maximum number of tasks, across all jobs in the queue, \n    which can be initialized concurrently. Once the queue's jobs exceed this \n    limit they will be queued on disk.  \n    ",
+        "property_value" : "200000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.maximum-initialized-active-tasks",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_local_dir",
+      "StackConfigurations" : {
+        "property_description" : "MapRed Local Directories.",
+        "property_value" : "/hadoop/mapred",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred_local_dir",
+        "filename" : "global.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.inmem.merge.threshold",
+      "StackConfigurations" : {
+        "property_description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
+        "property_value" : "1000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.inmem.merge.threshold",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.queue.default.acl-administer-jobs",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "*",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.queue.default.acl-administer-jobs",
+        "filename" : "mapred-queue-acls.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.reduce.input.limit",
+      "StackConfigurations" : {
+        "property_description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
+        "property_value" : "10737418240",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.reduce.input.limit",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.cluster.administrators",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : " hadoop",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.cluster.administrators",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.healthChecker.script.timeout",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "60000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.healthChecker.script.timeout",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.parallel.copies",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : "30",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.reduce.parallel.copies",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.tracker.history.completed.location",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : "/mapred/history/done",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.tracker.history.completed.location",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.reduce.slowstart.completed.maps",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "0.05",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.reduce.slowstart.completed.maps",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.tasktracker.tasks.sleeptime-before-sigkill",
+      "StackConfigurations" : {
+        "property_description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
+        "property_value" : "250",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.capacity",
+      "StackConfigurations" : {
+        "property_description" : "Percentage of the number of slots in the cluster that are\n      to be available for jobs in this queue.\n    ",
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.capacity",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.job.shuffle.input.buffer.percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
+        "property_value" : "0.7",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.job.shuffle.input.buffer.percent",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.child.root.logger",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "INFO,TLA",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.child.root.logger",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.fileoutputcommitter.marksuccessfuljobs",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-init-accept-jobs-factor",
+      "StackConfigurations" : {
+        "property_description" : "The default multipe of (maximum-system-jobs * queue-capacity) \n    used to determine the number of jobs which are accepted by the scheduler.  \n    ",
+        "property_value" : "10",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-init-accept-jobs-factor",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred_system_dir",
+      "StackConfigurations" : {
+        "property_description" : "MapRed System Directories.",
+        "property_value" : "/mapred/system",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred_system_dir",
+        "filename" : "global.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/io.sort.factor",
+      "StackConfigurations" : {
+        "property_description" : "No description",
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "io.sort.factor",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/jtnode_opt_maxnewsize",
+      "StackConfigurations" : {
+        "property_description" : "MapRed Capacity Scheduler.",
+        "property_value" : "200",
+        "stack_version" : "1.3.0",
+        "property_name" : "jtnode_opt_maxnewsize",
+        "filename" : "global.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.user-limit-factor",
+      "StackConfigurations" : {
+        "property_description" : "The multiple of the queue capacity which can be configured to \n    allow a single user to acquire more slots. \n    ",
+        "property_value" : "1",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.user-limit-factor",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.maximum-system-jobs",
+      "StackConfigurations" : {
+        "property_description" : "Maximum number of jobs in the system which can be initialized,\n     concurrently, by the CapacityScheduler.\n    ",
+        "property_value" : "3000",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.maximum-system-jobs",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.default-minimum-user-limit-percent",
+      "StackConfigurations" : {
+        "property_description" : "The percentage of the resources limited to a particular user\n      for the job queue at any given point of time by default.\n    ",
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.default-minimum-user-limit-percent",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.queue.default.minimum-user-limit-percent",
+      "StackConfigurations" : {
+        "property_description" : " Each queue enforces a limit on the percentage of resources \n    allocated to a user at any given time, if there is competition for them. \n    This user limit can vary between a minimum and maximum value. The former\n    depends on the number of users who have submitted jobs, and the latter is\n    set to this property value. For example, suppose the value of this \n    property is 25. If two users have submitted jobs to a queue, no single \n    user can use more than 50% of the queue resources. If a third user submits\n    a job, no single user can use more than 33% of the queue resources. With 4 \n    or more users, no user can use more than 25% of the queue's resources. A \n    value of 100 implies no user limits are imposed. \n    ",
+        "property_value" : "100",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.queue.default.minimum-user-limit-percent",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapreduce.jobtracker.staging.root.dir",
+      "StackConfigurations" : {
+        "property_description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n   name. It is a path in the default file system.",
+        "property_value" : "/user",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapreduce.jobtracker.staging.root.dir",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/mapred.capacity-scheduler.init-worker-threads",
+      "StackConfigurations" : {
+        "property_description" : "Number of worker threads which would be used by\n    Initialization poller to initialize jobs in a set of queue.\n    If number mentioned in property is equal to number of job queues\n    then a single thread would initialize jobs in a queue. If lesser\n    then a thread would get a set of queues assigned. If the number\n    is greater then number of threads would be equal to number of \n    job queues.\n    ",
+        "property_value" : "5",
+        "stack_version" : "1.3.0",
+        "property_name" : "mapred.capacity-scheduler.init-worker-threads",
+        "filename" : "capacity-scheduler.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE/configurations/tasktracker.http.threads",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "50",
+        "stack_version" : "1.3.0",
+        "property_name" : "tasktracker.http.threads",
+        "filename" : "mapred-site.xml",
+        "service_name" : "MAPREDUCE",
+        "stack_name" : "HDP"
+      }
+    }
+  ]
+}

+ 316 - 154
ambari-web/app/assets/data/wizard/stack/hdp/version130/OOZIE.json

@@ -1,155 +1,317 @@
 {
 {
-  "name" : "OOZIE",
-  "version" : "3.2.0",
-  "user" : "root",
-  "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-  "properties" : [ {
-    "name" : "oozie.base.url",
-    "value" : "http://localhost:11000/oozie",
-    "description" : "Base Oozie URL.",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.system.id",
-    "value" : "oozie-${user.name}",
-    "description" : "\n    The Oozie system ID.\n    ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.systemmode",
-    "value" : "NORMAL",
-    "description" : "\n     System mode for  Oozie at startup.\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.AuthorizationService.security.enabled",
-    "value" : "true",
-    "description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.PurgeService.older.than",
-    "value" : "30",
-    "description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.PurgeService.purge.interval",
-    "value" : "3600",
-    "description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.CallableQueueService.queue.size",
-    "value" : "1000",
-    "description" : "Max callable queue size",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.CallableQueueService.threads",
-    "value" : "10",
-    "description" : "Number of threads used for executing callables",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.CallableQueueService.callable.concurrency",
-    "value" : "3",
-    "description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.coord.normal.default.timeout",
-    "value" : "120",
-    "description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.db.schema.name",
-    "value" : "oozie",
-    "description" : "\n      Oozie DataBase Name\n     ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
-    "value" : " ",
-    "description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.authentication.type",
-    "value" : "simple",
-    "description" : "\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
-    "value" : " ",
-    "description" : "\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.WorkflowAppService.system.libpath",
-    "value" : "/user/${user.name}/share/lib",
-    "description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
-    "value" : "false",
-    "description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.authentication.kerberos.name.rules",
-    "value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
-    "description" : "The mapping from kerberos principal names to local OS user names.",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
-    "value" : "*=/etc/hadoop/conf",
-    "description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.ActionService.executor.ext.classes",
-    "value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
-    "description" : null,
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.SchemaService.wf.ext.schemas",
-    "value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
-    "description" : null,
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.create.db.schema",
-    "value" : "false",
-    "description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.jdbc.driver",
-    "value" : "org.apache.derby.jdbc.EmbeddedDriver",
-    "description" : "\n            JDBC driver class.\n        ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.jdbc.url",
-    "value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
-    "description" : "\n            JDBC URL.\n        ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.jdbc.username",
-    "value" : "sa",
-    "description" : "\n            DB user name.\n        ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.jdbc.password",
-    "value" : " ",
-    "description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
-    "filename" : "oozie-site.xml"
-  }, {
-    "name" : "oozie.service.JPAService.pool.max.active.conn",
-    "value" : "10",
-    "description" : "\n             Max number of connections.\n        ",
-    "filename" : "oozie-site.xml"
-  } ],
-  "components" : [ {
-    "name" : "OOZIE_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "OOZIE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "OOZIE_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.systemmode",
+      "StackConfigurations" : {
+        "property_description" : "\n     System mode for  Oozie at startup.\n     ",
+        "property_value" : "NORMAL",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.systemmode",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.kerberos.name.rules",
+      "StackConfigurations" : {
+        "property_description" : "The mapping from kerberos principal names to local OS user names.",
+        "property_value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.authentication.kerberos.name.rules",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.AuthorizationService.security.enabled",
+      "StackConfigurations" : {
+        "property_description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
+        "property_value" : "true",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.AuthorizationService.security.enabled",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.threads",
+      "StackConfigurations" : {
+        "property_description" : "Number of threads used for executing callables",
+        "property_value" : "10",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.CallableQueueService.threads",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.jobTracker.whitelist",
+      "StackConfigurations" : {
+        "property_description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
+        "property_value" : " ",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.pool.max.active.conn",
+      "StackConfigurations" : {
+        "property_description" : "\n             Max number of connections.\n        ",
+        "property_value" : "10",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.pool.max.active.conn",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.base.url",
+      "StackConfigurations" : {
+        "property_description" : "Base Oozie URL.",
+        "property_value" : "http://localhost:11000/oozie",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.base.url",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.SchemaService.wf.ext.schemas",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.SchemaService.wf.ext.schemas",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.older.than",
+      "StackConfigurations" : {
+        "property_description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
+        "property_value" : "30",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.PurgeService.older.than",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.nameNode.whitelist",
+      "StackConfigurations" : {
+        "property_description" : "\n      ",
+        "property_value" : " ",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.coord.normal.default.timeout",
+      "StackConfigurations" : {
+        "property_description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
+        "property_value" : "120",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.coord.normal.default.timeout",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/use.system.libpath.for.mapreduce.and.pig.jobs",
+      "StackConfigurations" : {
+        "property_description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.queue.size",
+      "StackConfigurations" : {
+        "property_description" : "Max callable queue size",
+        "property_value" : "1000",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.CallableQueueService.queue.size",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.ActionService.executor.ext.classes",
+      "StackConfigurations" : {
+        "property_description" : null,
+        "property_value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.ActionService.executor.ext.classes",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.password",
+      "StackConfigurations" : {
+        "property_description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
+        "property_value" : " ",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.jdbc.password",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.WorkflowAppService.system.libpath",
+      "StackConfigurations" : {
+        "property_description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
+        "property_value" : "/user/${user.name}/share/lib",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.WorkflowAppService.system.libpath",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.driver",
+      "StackConfigurations" : {
+        "property_description" : "\n            JDBC driver class.\n        ",
+        "property_value" : "org.apache.derby.jdbc.EmbeddedDriver",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.jdbc.driver",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.HadoopAccessorService.hadoop.configurations",
+      "StackConfigurations" : {
+        "property_description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
+        "property_value" : "*=/etc/hadoop/conf",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.PurgeService.purge.interval",
+      "StackConfigurations" : {
+        "property_description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
+        "property_value" : "3600",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.PurgeService.purge.interval",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.create.db.schema",
+      "StackConfigurations" : {
+        "property_description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.create.db.schema",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.system.id",
+      "StackConfigurations" : {
+        "property_description" : "\n    The Oozie system ID.\n    ",
+        "property_value" : "oozie-${user.name}",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.system.id",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.username",
+      "StackConfigurations" : {
+        "property_description" : "\n            DB user name.\n        ",
+        "property_value" : "sa",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.jdbc.username",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.db.schema.name",
+      "StackConfigurations" : {
+        "property_description" : "\n      Oozie DataBase Name\n     ",
+        "property_value" : "oozie",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.db.schema.name",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.authentication.type",
+      "StackConfigurations" : {
+        "property_description" : "\n      ",
+        "property_value" : "simple",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.authentication.type",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.JPAService.jdbc.url",
+      "StackConfigurations" : {
+        "property_description" : "\n            JDBC URL.\n        ",
+        "property_value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.JPAService.jdbc.url",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE/configurations/oozie.service.CallableQueueService.callable.concurrency",
+      "StackConfigurations" : {
+        "property_description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
+        "property_value" : "3",
+        "stack_version" : "1.3.0",
+        "property_name" : "oozie.service.CallableQueueService.callable.concurrency",
+        "filename" : "oozie-site.xml",
+        "service_name" : "OOZIE",
+        "stack_name" : "HDP"
+      }
+    }
+  ]
+}

+ 172 - 89
ambari-web/app/assets/data/wizard/stack/hdp/version130/WEBHCAT.json

@@ -1,90 +1,173 @@
 {
 {
-  "name" : "WEBHCAT",
-  "version" : "0.5.0",
-  "user" : "root",
-  "comment" : "This is comment for WEBHCAT service",
-  "properties" : [ {
-    "name" : "templeton.port",
-    "value" : "50111",
-    "description" : "The HTTP port for the main server.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.hadoop.conf.dir",
-    "value" : "/etc/hadoop/conf",
-    "description" : "The path to the Hadoop configuration.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.jar",
-    "value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
-    "description" : "The path to the Templeton jar file.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.libjars",
-    "value" : "/usr/lib/zookeeper/zookeeper.jar",
-    "description" : "Jars to add the the classpath.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.hadoop",
-    "value" : "/usr/bin/hadoop",
-    "description" : "The path to the Hadoop executable.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.pig.archive",
-    "value" : "hdfs:///apps/webhcat/pig.tar.gz",
-    "description" : "The path to the Pig archive.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.pig.path",
-    "value" : "pig.tar.gz/pig/bin/pig",
-    "description" : "The path to the Pig executable.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.hcat",
-    "value" : "/usr/bin/hcat",
-    "description" : "The path to the hcatalog executable.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.hive.archive",
-    "value" : "hdfs:///apps/webhcat/hive.tar.gz",
-    "description" : "The path to the Hive archive.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.hive.path",
-    "value" : "hive.tar.gz/hive/bin/hive",
-    "description" : "The path to the Hive executable.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.storage.class",
-    "value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
-    "description" : "The class to use as storage",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.override.enabled",
-    "value" : "false",
-    "description" : "\n     Enable the override path in templeton.override.jars\n   ",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.streaming.jar",
-    "value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
-    "description" : "The hdfs path to the Hadoop streaming jar file.",
-    "filename" : "webhcat-site.xml"
-  }, {
-    "name" : "templeton.exec.timeout",
-    "value" : "60000",
-    "description" : "Time out for templeton api",
-    "filename" : "webhcat-site.xml"
-  } ],
-  "components" : [ {
-    "name" : "WEBHCAT_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "WEBHCAT_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations?fields=*",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.streaming.jar",
+      "StackConfigurations" : {
+        "property_description" : "The hdfs path to the Hadoop streaming jar file.",
+        "property_value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.streaming.jar",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.override.enabled",
+      "StackConfigurations" : {
+        "property_description" : "\n     Enable the override path in templeton.override.jars\n   ",
+        "property_value" : "false",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.override.enabled",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.archive",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Pig archive.",
+        "property_value" : "hdfs:///apps/webhcat/pig.tar.gz",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.pig.archive",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Hadoop executable.",
+        "property_value" : "/usr/bin/hadoop",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.hadoop",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hcat",
+      "StackConfigurations" : {
+        "property_description" : "The path to the hcatalog executable.",
+        "property_value" : "/usr/bin/hcat",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.hcat",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.exec.timeout",
+      "StackConfigurations" : {
+        "property_description" : "Time out for templeton api",
+        "property_value" : "60000",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.exec.timeout",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hadoop.conf.dir",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Hadoop configuration.",
+        "property_value" : "/etc/hadoop/conf",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.hadoop.conf.dir",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.jar",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Templeton jar file.",
+        "property_value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.jar",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.libjars",
+      "StackConfigurations" : {
+        "property_description" : "Jars to add the the classpath.",
+        "property_value" : "/usr/lib/zookeeper/zookeeper.jar",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.libjars",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.port",
+      "StackConfigurations" : {
+        "property_description" : "The HTTP port for the main server.",
+        "property_value" : "50111",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.port",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.storage.class",
+      "StackConfigurations" : {
+        "property_description" : "The class to use as storage",
+        "property_value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.storage.class",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.archive",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Hive archive.",
+        "property_value" : "hdfs:///apps/webhcat/hive.tar.gz",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.hive.archive",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.hive.path",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Hive executable.",
+        "property_value" : "hive.tar.gz/hive/bin/hive",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.hive.path",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT/configurations/templeton.pig.path",
+      "StackConfigurations" : {
+        "property_description" : "The path to the Pig executable.",
+        "property_value" : "pig.tar.gz/pig/bin/pig",
+        "stack_version" : "1.3.0",
+        "property_name" : "templeton.pig.path",
+        "filename" : "webhcat-site.xml",
+        "service_name" : "WEBHCAT",
+        "stack_name" : "HDP"
+      }
+    }
+  ]
+}

+ 3 - 24
ambari-web/app/assets/data/wizard/stack/hdp/version130/ZOOKEEPER.json

@@ -1,25 +1,4 @@
 {
 {
-  "name" : "ZOOKEEPER",
-  "version" : "3.4.5",
-  "user" : "root",
-  "comment" : "This is comment for ZOOKEEPER service",
-  "properties" : [ ],
-  "components" : [ {
-    "name" : "ZOOKEEPER_SERVER",
-    "category" : "MASTER",
-    "client" : false,
-    "master" : true
-  }, {
-    "name" : "ZOOKEEPER_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  } ],
-  "clientOnlyService" : false,
-  "clientComponent" : {
-    "name" : "ZOOKEEPER_CLIENT",
-    "category" : "CLIENT",
-    "client" : true,
-    "master" : false
-  }
-}
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER/configurations?fields=*",
+  "items" : [ ]
+}

+ 584 - 4395
ambari-web/app/assets/data/wizard/stack/stacks.json

@@ -1,4395 +1,584 @@
-[ {
-  "name" : "HDP",
-  "version" : "1.2.2",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "1.3.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "HUE",
-    "version" : "2.2.0",
-    "user" : "root",
-    "comment" : "Hue is a graphical user interface to operate and develop\n      applications for Apache Hadoop.",
-    "components" : [ {
-      "name" : "HUE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HUE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "0.1",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-6&arch=$basearch"
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for GANGLIA service",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for OOZIE service",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.0",
-    "user" : "mapred",
-    "comment" : "This is comment for MAPREDUCE service",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HIVE service",
-    "components" : [ {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "TEMPLETON",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for TEMPLETON service",
-    "components" : [ {
-      "name" : "TEMPLETON_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TEMPLETON_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "TEMPLETON_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HDFS service",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for NAGIOS service",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for SQOOP service",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for PIG service",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "1.0",
-    "user" : "mapred",
-    "comment" : "This is comment for HBASE service",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "1.2.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.3.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.2",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDPLocal",
-  "version" : "1.2.2",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDPLocal",
-  "version" : "1.2.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.2",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "1.2.2",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "1.3.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.3.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.3.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "HUE",
-    "version" : "2.2.0",
-    "user" : "root",
-    "comment" : "Hue is a graphical user interface to operate and develop\n      applications for Apache Hadoop.",
-    "components" : [ {
-      "name" : "HUE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HUE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "0.1",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-1.1.1.16/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.1.1.16",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "epel",
-    "repoName" : "epel",
-    "mirrorsList" : "https://mirrors.fedoraproject.org/metalink?repo=epel-6&arch=$basearch"
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for GANGLIA service",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for OOZIE service",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.0",
-    "user" : "mapred",
-    "comment" : "This is comment for MAPREDUCE service",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HIVE service",
-    "components" : [ {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "TEMPLETON",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for TEMPLETON service",
-    "components" : [ {
-      "name" : "TEMPLETON_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TEMPLETON_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "TEMPLETON_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HDFS service",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for NAGIOS service",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for SQOOP service",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for PIG service",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "1.0",
-    "user" : "mapred",
-    "comment" : "This is comment for HBASE service",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "1.0",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  } ]
-}, {
-  "name" : "HDP",
-  "version" : "1.2.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.2",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDPLocal",
-  "version" : "1.2.2",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.2/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.2",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://public-repo-1.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.5",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-}, {
-  "name" : "HDPLocal",
-  "version" : "1.2.0",
-  "minUpgradeVersion" : null,
-  "repositories" : [ {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "centos6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "centos5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "centos5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos6",
-    "osType" : "redhat6",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat6",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-6&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/centos5",
-    "osType" : "redhat5",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : null,
-    "osType" : "redhat5",
-    "repoId" : "HDP-epel",
-    "repoName" : "HDP-epel",
-    "mirrorsList" : "http://mirrors.fedoraproject.org/mirrorlist?repo=epel-5&arch=$basearch"
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "suse11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-1.2.0/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-1.2.0",
-    "repoName" : "HDP",
-    "mirrorsList" : null
-  }, {
-    "baseUrl" : "http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.15/repos/suse11",
-    "osType" : "sles11",
-    "repoId" : "HDP-UTILS-1.1.0.15",
-    "repoName" : "HDP-UTILS",
-    "mirrorsList" : null
-  } ],
-  "services" : [ {
-    "name" : "GANGLIA",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "Ganglia Metrics Collection system",
-    "components" : [ {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "GANGLIA_MONITOR",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MONITOR_WEBSERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "GANGLIA_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "OOZIE",
-    "version" : "3.2.0",
-    "user" : "root",
-    "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
-    "components" : [ {
-      "name" : "OOZIE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "OOZIE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "MAPREDUCE",
-    "version" : "1.1.2",
-    "user" : "mapred",
-    "comment" : "Apache Hadoop Distributed Processing Framework",
-    "components" : [ {
-      "name" : "JOBTRACKER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "TASKTRACKER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "MAPREDUCE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HIVE",
-    "version" : "0.10.0",
-    "user" : "root",
-    "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
-    "components" : [ {
-      "name" : "HIVE_METASTORE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "MYSQL_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HIVE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HDFS",
-    "version" : "1.1.2",
-    "user" : "root",
-    "comment" : "Apache Hadoop Distributed File System",
-    "components" : [ {
-      "name" : "NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "DATANODE",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "SECONDARY_NAMENODE",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HDFS_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "NAGIOS",
-    "version" : "3.2.3",
-    "user" : "root",
-    "comment" : "Nagios Monitoring and Alerting system",
-    "components" : [ {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "NAGIOS_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  }, {
-    "name" : "SQOOP",
-    "version" : "1.4.2",
-    "user" : "root",
-    "comment" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases",
-    "components" : [ {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "SQOOP",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "PIG",
-    "version" : "0.10.1",
-    "user" : "root",
-    "comment" : "Scripting platform for analyzing large datasets",
-    "components" : [ {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "PIG",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HBASE",
-    "version" : "0.94.2",
-    "user" : "mapred",
-    "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
-    "components" : [ {
-      "name" : "HBASE_MASTER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "HBASE_REGIONSERVER",
-      "category" : "SLAVE",
-      "client" : false,
-      "master" : false
-    }, {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "HBASE_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "HCATALOG",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for HCATALOG service",
-    "components" : [ {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : true,
-    "clientComponent" : {
-      "name" : "HCAT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "ZOOKEEPER",
-    "version" : "3.4.5",
-    "user" : "root",
-    "comment" : "This is comment for ZOOKEEPER service",
-    "components" : [ {
-      "name" : "ZOOKEEPER_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }, {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "ZOOKEEPER_CLIENT",
-      "category" : "CLIENT",
-      "client" : true,
-      "master" : false
-    }
-  }, {
-    "name" : "WEBHCAT",
-    "version" : "0.5.0",
-    "user" : "root",
-    "comment" : "This is comment for WEBHCAT service",
-    "components" : [ {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    } ],
-    "clientOnlyService" : false,
-    "clientComponent" : {
-      "name" : "WEBHCAT_SERVER",
-      "category" : "MASTER",
-      "client" : false,
-      "master" : true
-    }
-  } ]
-} ]
+{
+  "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions?fields=stackServices/StackServices,Versions",
+  "items" : [
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0",
+      "Versions" : {
+        "stack_version" : "1.3.0",
+        "stack_name" : "HDP",
+        "min_upgrade_version" : "1.2.0"
+      },
+      "stackServices" : [
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/OOZIE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "OOZIE",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "System for workflow coordination and execution of Apache Hadoop jobs"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HCATALOG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "HCATALOG",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for HCATALOG service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/MAPREDUCE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.3.0",
+            "service_name" : "MAPREDUCE",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed Processing Framework"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HBASE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.3.0",
+            "service_name" : "HBASE",
+            "stack_name" : "HDP",
+            "service_version" : "0.94.5",
+            "comments" : "Non-relational distributed database and centralized service for configuration management & synchronization"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/ZOOKEEPER",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "ZOOKEEPER",
+            "stack_name" : "HDP",
+            "service_version" : "3.4.5",
+            "comments" : "This is comment for ZOOKEEPER service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HUE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "HUE",
+            "stack_name" : "HDP",
+            "service_version" : "2.2.0",
+            "comments" : "Hue is a graphical user interface to operate and develop\n      applications for Apache Hadoop."
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HDFS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "HDFS",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed File System"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/GANGLIA",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "GANGLIA",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "Ganglia Metrics Collection system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/WEBHCAT",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "WEBHCAT",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for WEBHCAT service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/SQOOP",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "SQOOP",
+            "stack_name" : "HDP",
+            "service_version" : "1.4.2",
+            "comments" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/NAGIOS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "NAGIOS",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.3",
+            "comments" : "Nagios Monitoring and Alerting system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/PIG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "PIG",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.1",
+            "comments" : "Scripting platform for analyzing large datasets"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.3.0/stackServices/HIVE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.3.0",
+            "service_name" : "HIVE",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.0",
+            "comments" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service"
+          }
+        }
+      ]
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1",
+      "Versions" : {
+        "stack_version" : "1.2.1",
+        "stack_name" : "HDP",
+        "min_upgrade_version" : null
+      },
+      "stackServices" : [
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/HIVE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "HIVE",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.0",
+            "comments" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/HBASE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.2.1",
+            "service_name" : "HBASE",
+            "stack_name" : "HDP",
+            "service_version" : "0.94.5",
+            "comments" : "Non-relational distributed database and centralized service for configuration management & synchronization"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/ZOOKEEPER",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "ZOOKEEPER",
+            "stack_name" : "HDP",
+            "service_version" : "3.4.5",
+            "comments" : "This is comment for ZOOKEEPER service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/HCATALOG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "HCATALOG",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for HCATALOG service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/HDFS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "HDFS",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed File System"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/NAGIOS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "NAGIOS",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.3",
+            "comments" : "Nagios Monitoring and Alerting system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/OOZIE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "OOZIE",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "System for workflow coordination and execution of Apache Hadoop jobs"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/GANGLIA",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "GANGLIA",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "Ganglia Metrics Collection system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/PIG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "PIG",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.1",
+            "comments" : "Scripting platform for analyzing large datasets"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/MAPREDUCE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.2.1",
+            "service_name" : "MAPREDUCE",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed Processing Framework"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/WEBHCAT",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "WEBHCAT",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for WEBHCAT service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.1/stackServices/SQOOP",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.1",
+            "service_name" : "SQOOP",
+            "stack_name" : "HDP",
+            "service_version" : "1.4.2",
+            "comments" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases"
+          }
+        }
+      ]
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0",
+      "Versions" : {
+        "stack_version" : "1.2.0",
+        "stack_name" : "HDP",
+        "min_upgrade_version" : null
+      },
+      "stackServices" : [
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/PIG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "PIG",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.1",
+            "comments" : "Scripting platform for analyzing large datasets"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/SQOOP",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "SQOOP",
+            "stack_name" : "HDP",
+            "service_version" : "1.4.2",
+            "comments" : "Tool for transferring bulk data between Apache Hadoop and structured data stores such as relational databases"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/GANGLIA",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "GANGLIA",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "Ganglia Metrics Collection system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/ZOOKEEPER",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "ZOOKEEPER",
+            "stack_name" : "HDP",
+            "service_version" : "3.4.5",
+            "comments" : "This is comment for ZOOKEEPER service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/NAGIOS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "NAGIOS",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.3",
+            "comments" : "Nagios Monitoring and Alerting system"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/HCATALOG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "HCATALOG",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for HCATALOG service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/WEBHCAT",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "WEBHCAT",
+            "stack_name" : "HDP",
+            "service_version" : "0.5.0",
+            "comments" : "This is comment for WEBHCAT service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/HIVE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "HIVE",
+            "stack_name" : "HDP",
+            "service_version" : "0.10.0",
+            "comments" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/HBASE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.2.0",
+            "service_name" : "HBASE",
+            "stack_name" : "HDP",
+            "service_version" : "0.94.2",
+            "comments" : "Non-relational distributed database and centralized service for configuration management & synchronization"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/MAPREDUCE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "1.2.0",
+            "service_name" : "MAPREDUCE",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed Processing Framework"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/HDFS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "HDFS",
+            "stack_name" : "HDP",
+            "service_version" : "1.1.2",
+            "comments" : "Apache Hadoop Distributed File System"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/1.2.0/stackServices/OOZIE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "1.2.0",
+            "service_name" : "OOZIE",
+            "stack_name" : "HDP",
+            "service_version" : "3.2.0",
+            "comments" : "System for workflow coordination and execution of Apache Hadoop jobs"
+          }
+        }
+      ]
+    },
+    {
+      "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1",
+      "Versions" : {
+        "stack_version" : "0.1",
+        "stack_name" : "HDP",
+        "min_upgrade_version" : null
+      },
+      "stackServices" : [
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/GANGLIA",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "GANGLIA",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for GANGLIA service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/HIVE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "HIVE",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for HIVE service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/TEMPLETON",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "TEMPLETON",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for TEMPLETON service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/OOZIE",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "OOZIE",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for OOZIE service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/ZOOKEEPER",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "ZOOKEEPER",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for ZOOKEEPER service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/SQOOP",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "SQOOP",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for SQOOP service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/MAPREDUCE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "0.1",
+            "service_name" : "MAPREDUCE",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for MAPREDUCE service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/HCATALOG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "HCATALOG",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for HCATALOG service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/HDFS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "HDFS",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for HDFS service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/HBASE",
+          "StackServices" : {
+            "user_name" : "mapred",
+            "stack_version" : "0.1",
+            "service_name" : "HBASE",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for HBASE service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/PIG",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "PIG",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for PIG service"
+          }
+        },
+        {
+          "href" : "http://dev.hortonworks.com:8080/api/v1/stacks2/HDP/versions/0.1/stackServices/NAGIOS",
+          "StackServices" : {
+            "user_name" : "root",
+            "stack_version" : "0.1",
+            "service_name" : "NAGIOS",
+            "stack_name" : "HDP",
+            "service_version" : "1.0",
+            "comments" : "This is comment for NAGIOS service"
+          }
+        }
+      ]
+    }
+  ]
+}

+ 37 - 34
ambari-web/app/controllers/main/admin/cluster.js

@@ -28,13 +28,13 @@ App.MainAdminClusterController = Em.Controller.extend({
   updateUpgradeVersion: function(){
   updateUpgradeVersion: function(){
     if(App.router.get('clusterController.isLoaded')){
     if(App.router.get('clusterController.isLoaded')){
       var url = App.formatUrl(
       var url = App.formatUrl(
-        App.apiPrefix + "/stacks",
+        App.apiPrefix + "/stacks2/HDP/versions?fields=stackServices/StackServices,Versions",
         {},
         {},
         '/data/wizard/stack/stacks.json'
         '/data/wizard/stack/stacks.json'
       );
       );
       var upgradeVersion = this.get('upgradeVersion') || App.defaultStackVersion;
       var upgradeVersion = this.get('upgradeVersion') || App.defaultStackVersion;
-      var installedServices = {};
-      var newServices = {};
+      var currentStack = {};
+      var upgradeStack = {};
       $.ajax({
       $.ajax({
         type: "GET",
         type: "GET",
         url: url,
         url: url,
@@ -45,17 +45,16 @@ App.MainAdminClusterController = Em.Controller.extend({
           var currentVersion = App.currentStackVersion.replace(/HDP-/, '');
           var currentVersion = App.currentStackVersion.replace(/HDP-/, '');
           var minUpgradeVersion = currentVersion;
           var minUpgradeVersion = currentVersion;
           upgradeVersion = upgradeVersion.replace(/HDP-/, '');
           upgradeVersion = upgradeVersion.replace(/HDP-/, '');
-          data = data.filterProperty('name', 'HDP');
-          data.mapProperty('version').forEach(function(version){
+          data.items.mapProperty('Versions.stack_version').forEach(function(version){
             upgradeVersion = (upgradeVersion < version) ? version : upgradeVersion;
             upgradeVersion = (upgradeVersion < version) ? version : upgradeVersion;
           });
           });
-          //TODO remove hardcoded upgrade version
-          upgradeVersion = (App.testMode)?'1.3.0': upgradeVersion;
-          minUpgradeVersion = data.findProperty('version', upgradeVersion).minUpgradeVersion;
-          minUpgradeVersion = (minUpgradeVersion) ? minUpgradeVersion : currentVersion;
-          upgradeVersion = (minUpgradeVersion <= currentVersion) ? upgradeVersion : currentVersion;
-          installedServices = data.findProperty('version', currentVersion);
-          newServices = data.findProperty('version', upgradeVersion);
+          currentStack = data.items.findProperty('Versions.stack_version', currentVersion);
+          upgradeStack = data.items.findProperty('Versions.stack_version', upgradeVersion);
+          minUpgradeVersion = upgradeStack.Versions.min_upgrade_version;
+          if(minUpgradeVersion && (minUpgradeVersion > currentVersion)){
+            upgradeVersion = currentVersion;
+            upgradeStack = currentStack;
+          }
           upgradeVersion = 'HDP-' + upgradeVersion;
           upgradeVersion = 'HDP-' + upgradeVersion;
         },
         },
         error: function (request, ajaxOptions, error) {
         error: function (request, ajaxOptions, error) {
@@ -64,8 +63,8 @@ App.MainAdminClusterController = Em.Controller.extend({
         statusCode: require('data/statusCodes')
         statusCode: require('data/statusCodes')
       });
       });
       this.set('upgradeVersion', upgradeVersion);
       this.set('upgradeVersion', upgradeVersion);
-      if(installedServices && newServices){
-        this.parseServicesInfo(installedServices, newServices);
+      if(currentStack && upgradeStack){
+        this.parseServicesInfo(currentStack, upgradeStack);
       } else {
       } else {
         console.log('HDP stack doesn\'t have services with defaultStackVersion');
         console.log('HDP stack doesn\'t have services with defaultStackVersion');
       }
       }
@@ -74,32 +73,36 @@ App.MainAdminClusterController = Em.Controller.extend({
   /**
   /**
    * parse services info(versions, description) by version
    * parse services info(versions, description) by version
    */
    */
-  parseServicesInfo: function (oldServices, newServices) {
+  parseServicesInfo: function (currentStack, upgradeStack) {
     var result = [];
     var result = [];
     var installedServices = App.Service.find().mapProperty('serviceName');
     var installedServices = App.Service.find().mapProperty('serviceName');
     var displayOrderConfig = require('data/services');
     var displayOrderConfig = require('data/services');
-    if(oldServices.services && newServices.services){
+    if(currentStack.stackServices.length && upgradeStack.stackServices.length){
       // loop through all the service components
       // loop through all the service components
       for (var i = 0; i < displayOrderConfig.length; i++) {
       for (var i = 0; i < displayOrderConfig.length; i++) {
-        var entry = oldServices.services.findProperty("name", displayOrderConfig[i].serviceName);
+        var entry = currentStack.stackServices.
+          findProperty("StackServices.service_name", displayOrderConfig[i].serviceName);
         if (entry) {
         if (entry) {
-          if (installedServices.contains(entry.name)) {
-            var myService = Em.Object.create({
-              serviceName: entry.name,
-              displayName: displayOrderConfig[i].displayName,
-              isDisabled: i === 0,
-              isSelected: true,
-              isInstalled: false,
-              isHidden: displayOrderConfig[i].isHidden,
-              description: entry.comment,
-              version: entry.version,
-              newVersion: newServices.services.findProperty("name", displayOrderConfig[i].serviceName).version
-            });
-            //From 1.3.0 for Hive we display only "Hive" (but it installes HCat and WebHCat as well)
-            if (this.get('upgradeVersion').replace(/HDP-/, '') >= '1.3.0' && displayOrderConfig[i].serviceName == 'HIVE') {
-              myService.set('displayName', 'Hive');
-            }
-            result.push(myService);
+          entry = entry.StackServices;
+        if (installedServices.contains(entry.service_name)) {
+          var myService = Em.Object.create({
+            serviceName: entry.service_name,
+            displayName: displayOrderConfig[i].displayName,
+            isDisabled: i === 0,
+            isSelected: true,
+            isInstalled: false,
+            isHidden: displayOrderConfig[i].isHidden,
+            description: entry.comments,
+            version: entry.service_version,
+            newVersion: upgradeStack.stackServices.
+              findProperty("StackServices.service_name", displayOrderConfig[i].serviceName).
+              StackServices.service_version
+          });
+          //From 1.3.0 for Hive we display only "Hive" (but it install HCat and WebHCat as well)
+          if (this.get('upgradeVersion').replace(/HDP-/, '') >= '1.3.0' && displayOrderConfig[i].serviceName == 'HIVE') {
+            myService.set('displayName', 'Hive');
+          }
+          result.push(myService);
           }
           }
         }
         }
         else {
         else {

+ 1 - 1
ambari-web/app/controllers/main/host/add_controller.js

@@ -93,7 +93,7 @@ App.AddHostController = App.WizardController.extend({
    */
    */
   loadServicesFromServer: function() {
   loadServicesFromServer: function() {
     var displayOrderConfig = require('data/services');
     var displayOrderConfig = require('data/services');
-    var apiUrl = App.get('stackVersionURL');
+    var apiUrl = App.get('stack2VersionURL');
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     //
     //
     apiService.forEach(function(item, index){
     apiService.forEach(function(item, index){

+ 1 - 1
ambari-web/app/controllers/main/service/add_controller.js

@@ -95,7 +95,7 @@ App.AddServiceController = App.WizardController.extend({
       return;
       return;
     }
     }
     var displayOrderConfig = require('data/services');
     var displayOrderConfig = require('data/services');
-    var apiUrl = App.get('stackVersionURL');
+    var apiUrl = App.get('stack2VersionURL');
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     //
     //
     apiService.forEach(function(item, index){
     apiService.forEach(function(item, index){

+ 1 - 1
ambari-web/app/controllers/main/service/reassign_controller.js

@@ -71,7 +71,7 @@ App.ReassignMasterController = App.WizardController.extend({
    */
    */
   loadServicesFromServer: function() {
   loadServicesFromServer: function() {
     var displayOrderConfig = require('data/services');
     var displayOrderConfig = require('data/services');
-    var apiUrl = App.get('stackVersionURL');
+    var apiUrl = App.get('stack2VersionURL');
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     //
     //
     apiService.forEach(function(item, index){
     apiService.forEach(function(item, index){

+ 7 - 6
ambari-web/app/controllers/wizard.js

@@ -431,7 +431,7 @@ App.WizardController = Em.Controller.extend({
     var result = null;
     var result = null;
     var method = 'GET';
     var method = 'GET';
     var testUrl = '/data/wizard/stack/hdp/version/1.3.0.json';
     var testUrl = '/data/wizard/stack/hdp/version/1.3.0.json';
-    var url = (App.testMode) ? testUrl : App.apiPrefix + apiUrl;
+    var url = (App.testMode) ? testUrl : App.apiPrefix + apiUrl + '?fields=stackServices/StackServices';
     $.ajax({
     $.ajax({
       type: method,
       type: method,
       url: url,
       url: url,
@@ -458,17 +458,18 @@ App.WizardController = Em.Controller.extend({
 
 
         // loop through all the service components
         // loop through all the service components
         for (var i = 0; i < displayOrderConfig.length; i++) {
         for (var i = 0; i < displayOrderConfig.length; i++) {
-          var entry = jsonData.services.findProperty("name", displayOrderConfig[i].serviceName);
+          var entry = jsonData.stackServices.findProperty("StackServices.service_name", displayOrderConfig[i].serviceName);
           if (entry) {
           if (entry) {
+            entry = entry.StackServices;
             var myService = Service.create({
             var myService = Service.create({
-              serviceName: entry.name,
+              serviceName: entry.service_name,
               displayName: displayOrderConfig[i].displayName,
               displayName: displayOrderConfig[i].displayName,
               isDisabled: i === 0,
               isDisabled: i === 0,
               isSelected: true,
               isSelected: true,
               isInstalled: false,
               isInstalled: false,
               isHidden: displayOrderConfig[i].isHidden,
               isHidden: displayOrderConfig[i].isHidden,
-              description: entry.comment,
-              version: entry.version
+              description: entry.comments,
+              version: entry.service_version
             });
             });
 
 
             data.push(myService);
             data.push(myService);
@@ -501,7 +502,7 @@ App.WizardController = Em.Controller.extend({
       return;
       return;
     }
     }
     var displayOrderConfig = require('data/services');
     var displayOrderConfig = require('data/services');
-    var apiUrl = App.get('stackVersionURL');
+    var apiUrl = App.get('stack2VersionURL');
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     var apiService = this.loadServiceComponents(displayOrderConfig, apiUrl);
     this.set('content.services', apiService);
     this.set('content.services', apiService);
     App.db.setService(apiService);
     App.db.setService(apiService);

+ 1 - 1
ambari-web/app/routes/stack_upgrade.js

@@ -42,7 +42,7 @@ module.exports = Em.Route.extend({
       });
       });
       var statuses = ['STOPPING_SERVICES', 'STACK_UPGRADING', 'STACK_UPGRADE_FAILED', 'STACK_UPGRADED'];
       var statuses = ['STOPPING_SERVICES', 'STACK_UPGRADING', 'STACK_UPGRADE_FAILED', 'STACK_UPGRADED'];
       var currentClusterStatus = App.clusterStatus.get('value');
       var currentClusterStatus = App.clusterStatus.get('value');
-      App.db.data = currentClusterStatus.localdb;
+      if (currentClusterStatus.localdb) App.db.data = currentClusterStatus.localdb;
       if (statuses.contains(currentClusterStatus.clusterState)) {
       if (statuses.contains(currentClusterStatus.clusterState)) {
         stackUpgradeController.setCurrentStep(3);
         stackUpgradeController.setCurrentStep(3);
       }
       }

+ 1 - 1
ambari-web/app/utils/ajax.js

@@ -199,7 +199,7 @@ var urls = {
     }
     }
   },
   },
   'config.advanced': {
   'config.advanced': {
-    'real': '{stackVersionUrl}/services/{serviceName}',
+    'real': '{stack2VersionUrl}/stackServices/{serviceName}/configurations?fields=*',
     'mock': '/data/wizard/stack/hdp/version130/{serviceName}.json',
     'mock': '/data/wizard/stack/hdp/version130/{serviceName}.json',
     'format': function(data){
     'format': function(data){
       return {
       return {

+ 16 - 4
ambari-web/app/utils/config.js

@@ -458,7 +458,7 @@ App.config = Em.Object.create({
       sender: this,
       sender: this,
       data: {
       data: {
         serviceName: serviceName,
         serviceName: serviceName,
-        stackVersionUrl: App.get('stackVersionURL')
+        stack2VersionUrl: App.get('stack2VersionURL')
       },
       },
       success: 'loadAdvancedConfigSuccess'
       success: 'loadAdvancedConfigSuccess'
     });
     });
@@ -468,10 +468,22 @@ App.config = Em.Object.create({
 
 
   loadAdvancedConfigSuccess: function (data, opt, params) {
   loadAdvancedConfigSuccess: function (data, opt, params) {
     console.log("TRACE: In success function for the loadAdvancedConfig; url is ", opt.url);
     console.log("TRACE: In success function for the loadAdvancedConfig; url is ", opt.url);
-    data.properties.setEach('serviceName', params.serviceName);
-    serviceComponents[params.serviceName] = data.properties;
+    var properties = [];
+    if (data.items.length) {
+      data.items.forEach(function (item) {
+        item = item.StackConfigurations;
+        properties.push({
+          serviceName: item.service_name,
+          name: item.property_name,
+          value: item.property_value,
+          description: item.property_description,
+          filename: item.filename
+        });
+      }, this);
+      serviceComponents[data.items[0].StackConfigurations.service_name] = properties;
+    }
   },
   },
-  
+
   /**
   /**
    * Determine the map which shows which services
    * Determine the map which shows which services
    * each global property effects.
    * each global property effects.