123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323 |
- #!/usr/bin/env python
- '''
- Licensed to the Apache Software Foundation (ASF) under one
- or more contributor license agreements. See the NOTICE file
- distributed with this work for additional information
- regarding copyright ownership. The ASF licenses this file
- to you under the Apache License, Version 2.0 (the
- "License"); you may not use this file except in compliance
- with the License. You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License.
- '''
- import ConfigParser
- import StringIO
- import json
- from NetUtil import NetUtil
- content = """
- [server]
- hostname=localhost
- url_port=8440
- secured_url_port=8441
- [agent]
- prefix=/tmp/ambari-agent
- tmp_dir=/tmp/ambari-agent/tmp
- data_cleanup_interval=86400
- data_cleanup_max_age=2592000
- data_cleanup_max_size_MB = 100
- ping_port=8670
- cache_dir=/var/lib/ambari-agent/cache
- [services]
- [python]
- custom_actions_dir = /var/lib/ambari-agent/resources/custom_actions
- [command]
- maxretries=2
- sleepBetweenRetries=1
- [security]
- keysdir=/tmp/ambari-agent
- server_crt=ca.crt
- passphrase_env_var_name=AMBARI_PASSPHRASE
- [heartbeat]
- state_interval = 6
- dirs=/etc/hadoop,/etc/hadoop/conf,/var/run/hadoop,/var/log/hadoop
- rpms=glusterfs,openssl,wget,net-snmp,ntpd,ganglia,nagios,glusterfs
- log_lines_count=300
- """
- imports = [
- "hdp/manifests/*.pp",
- "hdp-hadoop/manifests/*.pp",
- "hdp-hbase/manifests/*.pp",
- "hdp-zookeeper/manifests/*.pp",
- "hdp-oozie/manifests/*.pp",
- "hdp-pig/manifests/*.pp",
- "hdp-sqoop/manifests/*.pp",
- "hdp-templeton/manifests/*.pp",
- "hdp-hive/manifests/*.pp",
- "hdp-hcat/manifests/*.pp",
- "hdp-mysql/manifests/*.pp",
- "hdp-monitor-webserver/manifests/*.pp",
- "hdp-repos/manifests/*.pp"
- ]
- rolesToClass = {
- 'GLUSTERFS': 'hdp-hadoop::glusterfs',
- 'GLUSTERFS_CLIENT': 'hdp-hadoop::glusterfs_client',
- 'GLUSTERFS_SERVICE_CHECK': 'hdp-hadoop::glusterfs_service_check',
- 'NAMENODE': 'hdp-hadoop::namenode',
- 'DATANODE': 'hdp-hadoop::datanode',
- 'SECONDARY_NAMENODE': 'hdp-hadoop::snamenode',
- 'JOBTRACKER': 'hdp-hadoop::jobtracker',
- 'TASKTRACKER': 'hdp-hadoop::tasktracker',
- 'RESOURCEMANAGER': 'hdp-yarn::resourcemanager',
- 'NODEMANAGER': 'hdp-yarn::nodemanager',
- 'HISTORYSERVER': 'hdp-yarn::historyserver',
- 'YARN_CLIENT': 'hdp-yarn::yarn_client',
- 'HDFS_CLIENT': 'hdp-hadoop::client',
- 'MAPREDUCE_CLIENT': 'hdp-hadoop::client',
- 'MAPREDUCE2_CLIENT': 'hdp-yarn::mapreducev2_client',
- 'ZOOKEEPER_SERVER': 'hdp-zookeeper',
- 'ZOOKEEPER_CLIENT': 'hdp-zookeeper::client',
- 'HBASE_MASTER': 'hdp-hbase::master',
- 'HBASE_REGIONSERVER': 'hdp-hbase::regionserver',
- 'HBASE_CLIENT': 'hdp-hbase::client',
- 'PIG': 'hdp-pig',
- 'SQOOP': 'hdp-sqoop',
- 'OOZIE_SERVER': 'hdp-oozie::server',
- 'OOZIE_CLIENT': 'hdp-oozie::client',
- 'HIVE_CLIENT': 'hdp-hive::client',
- 'HCAT': 'hdp-hcat',
- 'HIVE_SERVER': 'hdp-hive::server',
- 'HIVE_METASTORE': 'hdp-hive::metastore',
- 'MYSQL_SERVER': 'hdp-mysql::server',
- 'WEBHCAT_SERVER': 'hdp-templeton::server',
- 'DASHBOARD': 'hdp-dashboard',
- 'NAGIOS_SERVER': 'hdp-nagios::server',
- 'GANGLIA_SERVER': 'hdp-ganglia::server',
- 'GANGLIA_MONITOR': 'hdp-ganglia::monitor',
- 'HTTPD': 'hdp-monitor-webserver',
- 'HUE_SERVER': 'hdp-hue::server',
- 'HDFS_SERVICE_CHECK': 'hdp-hadoop::hdfs::service_check',
- 'MAPREDUCE_SERVICE_CHECK': 'hdp-hadoop::mapred::service_check',
- 'MAPREDUCE2_SERVICE_CHECK': 'hdp-yarn::mapred2::service_check',
- 'ZOOKEEPER_SERVICE_CHECK': 'hdp-zookeeper::zookeeper::service_check',
- 'ZOOKEEPER_QUORUM_SERVICE_CHECK': 'hdp-zookeeper::quorum::service_check',
- 'HBASE_SERVICE_CHECK': 'hdp-hbase::hbase::service_check',
- 'HIVE_SERVICE_CHECK': 'hdp-hive::hive::service_check',
- 'HCAT_SERVICE_CHECK': 'hdp-hcat::hcat::service_check',
- 'OOZIE_SERVICE_CHECK': 'hdp-oozie::oozie::service_check',
- 'PIG_SERVICE_CHECK': 'hdp-pig::pig::service_check',
- 'SQOOP_SERVICE_CHECK': 'hdp-sqoop::sqoop::service_check',
- 'WEBHCAT_SERVICE_CHECK': 'hdp-templeton::templeton::service_check',
- 'DASHBOARD_SERVICE_CHECK': 'hdp-dashboard::dashboard::service_check',
- 'DECOMMISSION_DATANODE': 'hdp-hadoop::hdfs::decommission',
- 'HUE_SERVICE_CHECK': 'hdp-hue::service_check',
- 'RESOURCEMANAGER_SERVICE_CHECK': 'hdp-yarn::resourcemanager::service_check',
- 'HISTORYSERVER_SERVICE_CHECK': 'hdp-yarn::historyserver::service_check',
- 'TEZ_CLIENT': 'hdp-tez::tez_client',
- 'YARN_SERVICE_CHECK': 'hdp-yarn::yarn::service_check',
- 'FLUME_SERVER': 'hdp-flume',
- 'JOURNALNODE': 'hdp-hadoop::journalnode',
- 'ZKFC': 'hdp-hadoop::zkfc'
- }
- serviceStates = {
- 'START': 'running',
- 'INSTALL': 'installed_and_configured',
- 'STOP': 'stopped'
- }
- servicesToPidNames = {
- 'GLUSTERFS' : 'glusterd.pid$',
- 'NAMENODE': 'hadoop-{USER}-namenode.pid$',
- 'SECONDARY_NAMENODE': 'hadoop-{USER}-secondarynamenode.pid$',
- 'DATANODE': 'hadoop-{USER}-datanode.pid$',
- 'JOBTRACKER': 'hadoop-{USER}-jobtracker.pid$',
- 'TASKTRACKER': 'hadoop-{USER}-tasktracker.pid$',
- 'RESOURCEMANAGER': 'yarn-{USER}-resourcemanager.pid$',
- 'NODEMANAGER': 'yarn-{USER}-nodemanager.pid$',
- 'HISTORYSERVER': 'mapred-{USER}-historyserver.pid$',
- 'JOURNALNODE': 'hadoop-{USER}-journalnode.pid$',
- 'ZKFC': 'hadoop-{USER}-zkfc.pid$',
- 'OOZIE_SERVER': 'oozie.pid',
- 'ZOOKEEPER_SERVER': 'zookeeper_server.pid',
- 'FLUME_SERVER': 'flume-node.pid',
- 'TEMPLETON_SERVER': 'templeton.pid',
- 'NAGIOS_SERVER': 'nagios.pid',
- 'GANGLIA_SERVER': 'gmetad.pid',
- 'GANGLIA_MONITOR': 'gmond.pid',
- 'HBASE_MASTER': 'hbase-{USER}-master.pid',
- 'HBASE_REGIONSERVER': 'hbase-{USER}-regionserver.pid',
- 'HCATALOG_SERVER': 'webhcat.pid',
- 'KERBEROS_SERVER': 'kadmind.pid',
- 'HIVE_SERVER': 'hive-server.pid',
- 'HIVE_METASTORE': 'hive.pid',
- 'MYSQL_SERVER': 'mysqld.pid',
- 'HUE_SERVER': '/var/run/hue/supervisor.pid',
- 'WEBHCAT_SERVER': 'webhcat.pid',
- }
- #Each service, which's pid depends on user should provide user mapping
- servicesToLinuxUser = {
- 'NAMENODE': 'hdfs_user',
- 'SECONDARY_NAMENODE': 'hdfs_user',
- 'DATANODE': 'hdfs_user',
- 'JOURNALNODE': 'hdfs_user',
- 'ZKFC': 'hdfs_user',
- 'JOBTRACKER': 'mapred_user',
- 'TASKTRACKER': 'mapred_user',
- 'RESOURCEMANAGER': 'yarn_user',
- 'NODEMANAGER': 'yarn_user',
- 'HISTORYSERVER': 'mapred_user',
- 'HBASE_MASTER': 'hbase_user',
- 'HBASE_REGIONSERVER': 'hbase_user',
- }
- pidPathesVars = [
- {'var' : 'glusterfs_pid_dir_prefix',
- 'defaultValue' : '/var/run'},
- {'var' : 'hadoop_pid_dir_prefix',
- 'defaultValue' : '/var/run/hadoop'},
- {'var' : 'hadoop_pid_dir_prefix',
- 'defaultValue' : '/var/run/hadoop'},
- {'var' : 'ganglia_runtime_dir',
- 'defaultValue' : '/var/run/ganglia/hdp'},
- {'var' : 'hbase_pid_dir',
- 'defaultValue' : '/var/run/hbase'},
- {'var' : '',
- 'defaultValue' : '/var/run/nagios'},
- {'var' : 'zk_pid_dir',
- 'defaultValue' : '/var/run/zookeeper'},
- {'var' : 'oozie_pid_dir',
- 'defaultValue' : '/var/run/oozie'},
- {'var' : 'hcat_pid_dir',
- 'defaultValue' : '/var/run/webhcat'},
- {'var' : 'hive_pid_dir',
- 'defaultValue' : '/var/run/hive'},
- {'var' : 'mysqld_pid_dir',
- 'defaultValue' : '/var/run/mysqld'},
- {'var' : 'hcat_pid_dir',
- 'defaultValue' : '/var/run/webhcat'},
- {'var' : 'yarn_pid_dir_prefix',
- 'defaultValue' : '/var/run/hadoop-yarn'},
- {'var' : 'mapred_pid_dir_prefix',
- 'defaultValue' : '/var/run/hadoop-mapreduce'},
- ]
- class AmbariConfig:
- TWO_WAY_SSL_PROPERTY = "security.server.two_way_ssl"
- CONFIG_FILE = "/etc/ambari-agent/conf/ambari-agent.ini"
- SERVER_CONNECTION_INFO = "{0}/connection_info"
- CONNECTION_PROTOCOL = "https"
- config = None
- net = None
- def __init__(self):
- global content
- self.config = ConfigParser.RawConfigParser()
- self.net = NetUtil()
- self.config.readfp(StringIO.StringIO(content))
- def get(self, section, value, default=None):
- try:
- return self.config.get(section, value)
- except ConfigParser.Error, err:
- if default:
- return default
- raise err
- def set(self, section, option, value):
- self.config.set(section, option, value)
- def add_section(self, section):
- self.config.add_section(section)
- def setConfig(self, customConfig):
- self.config = customConfig
- def getConfig(self):
- return self.config
- def getRolesToClass(self):
- global rolesToClass
- return rolesToClass
- def getServiceStates(self):
- global serviceStates
- return serviceStates
- def getServicesToPidNames(self):
- global servicesToPidNames
- return servicesToPidNames
- def getImports(self):
- global imports
- return imports
- def getPidPathesVars(self):
- global pidPathesVars
- return pidPathesVars
- def has_option(self, section, option):
- return self.config.has_option(section, option)
- def remove_option(self, section, option):
- return self.config.remove_option(section, option)
- def load(self, data):
- self.config = ConfigParser.RawConfigParser(data)
- def read(self, filename):
- self.config.read(filename)
- def getServerOption(self, url, name, default=None):
- status, response = self.net.checkURL(url)
- if status is True:
- try:
- data = json.loads(response)
- if name in data:
- return data[name]
- except:
- pass
- return default
- def get_api_url(self):
- return "%s://%s:%s" % (self.CONNECTION_PROTOCOL,
- self.get('server', 'hostname'),
- self.get('server', 'url_port'))
- def isTwoWaySSLConnection(self):
- req_url = self.get_api_url()
- response = self.getServerOption(self.SERVER_CONNECTION_INFO.format(req_url), self.TWO_WAY_SSL_PROPERTY, 'false')
- if response is None:
- return False
- elif response.lower() == "true":
- return True
- else:
- return False
- def main():
- print AmbariConfig().config
- if __name__ == "__main__":
- main()
|