|
@@ -50,15 +50,14 @@ class TestNamenode(RMFTestCase):
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
)
|
|
)
|
|
self.assert_configure_default()
|
|
self.assert_configure_default()
|
|
- self.assertResourceCalled('Execute', 'sudo ls /hadoop/hdfs/namenode | wc -l | grep -q ^0$',
|
|
|
|
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
|
|
|
|
- )
|
|
|
|
- self.assertResourceCalled('Execute', 'sudo su hdfs - -s /bin/bash -c "export PATH=$PATH:/usr/bin ; yes Y | hdfs --config /etc/hadoop/conf namenode -format"',
|
|
|
|
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
|
|
|
|
- )
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', 'ls /hadoop/hdfs/namenode | wc -l | grep -q ^0$',)
|
|
|
|
+ self.assertResourceCalled('Execute', 'yes Y | hdfs --config /etc/hadoop/conf namenode -format',
|
|
|
|
+ path = ['/usr/bin'],
|
|
|
|
+ user = 'hdfs',
|
|
|
|
+ )
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
- recursive = True,
|
|
|
|
- )
|
|
|
|
|
|
+ recursive = True,
|
|
|
|
+ )
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
owner = 'hdfs',
|
|
owner = 'hdfs',
|
|
content = Template('exclude_hosts_list.j2'),
|
|
content = Template('exclude_hosts_list.j2'),
|
|
@@ -81,7 +80,7 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
@@ -142,7 +141,7 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf stop namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf stop namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = None,
|
|
not_if = None,
|
|
)
|
|
)
|
|
@@ -171,15 +170,14 @@ class TestNamenode(RMFTestCase):
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
)
|
|
)
|
|
self.assert_configure_secured()
|
|
self.assert_configure_secured()
|
|
- self.assertResourceCalled('Execute', 'sudo ls /hadoop/hdfs/namenode | wc -l | grep -q ^0$',
|
|
|
|
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
|
|
|
|
- )
|
|
|
|
- self.assertResourceCalled('Execute', 'sudo su hdfs - -s /bin/bash -c "export PATH=$PATH:/usr/bin ; yes Y | hdfs --config /etc/hadoop/conf namenode -format"',
|
|
|
|
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
|
|
|
|
- )
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', 'ls /hadoop/hdfs/namenode | wc -l | grep -q ^0$',)
|
|
|
|
+ self.assertResourceCalled('Execute', 'yes Y | hdfs --config /etc/hadoop/conf namenode -format',
|
|
|
|
+ path = ['/usr/bin'],
|
|
|
|
+ user = 'hdfs',
|
|
|
|
+ )
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
- recursive = True,
|
|
|
|
- )
|
|
|
|
|
|
+ recursive = True,
|
|
|
|
+ )
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
owner = 'hdfs',
|
|
owner = 'hdfs',
|
|
content = Template('exclude_hosts_list.j2'),
|
|
content = Template('exclude_hosts_list.j2'),
|
|
@@ -202,7 +200,7 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
@@ -266,7 +264,7 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf stop namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf stop namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = None,
|
|
not_if = None,
|
|
)
|
|
)
|
|
@@ -306,14 +304,14 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
path = ['/usr/bin'],
|
|
path = ['/usr/bin'],
|
|
tries = 40,
|
|
tries = 40,
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
user = 'hdfs',
|
|
user = 'hdfs',
|
|
try_sleep = 10,
|
|
try_sleep = 10,
|
|
)
|
|
)
|
|
@@ -347,7 +345,7 @@ class TestNamenode(RMFTestCase):
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
action = ['create'],
|
|
action = ['create'],
|
|
bin_dir = '/usr/bin',
|
|
bin_dir = '/usr/bin',
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
)
|
|
)
|
|
self.assertNoMoreResources()
|
|
self.assertNoMoreResources()
|
|
|
|
|
|
@@ -382,7 +380,7 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
@@ -392,7 +390,7 @@ class TestNamenode(RMFTestCase):
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
path = ['/usr/bin'],
|
|
path = ['/usr/bin'],
|
|
tries = 40,
|
|
tries = 40,
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
user = 'hdfs',
|
|
user = 'hdfs',
|
|
try_sleep = 10,
|
|
try_sleep = 10,
|
|
)
|
|
)
|
|
@@ -426,7 +424,7 @@ class TestNamenode(RMFTestCase):
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
action = ['create'],
|
|
action = ['create'],
|
|
bin_dir = '/usr/bin',
|
|
bin_dir = '/usr/bin',
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
)
|
|
)
|
|
self.assertNoMoreResources()
|
|
self.assertNoMoreResources()
|
|
|
|
|
|
@@ -443,13 +441,13 @@ class TestNamenode(RMFTestCase):
|
|
self.assert_configure_default()
|
|
self.assert_configure_default()
|
|
|
|
|
|
# verify that active namenode was formatted
|
|
# verify that active namenode was formatted
|
|
- self.assertResourceCalled('Execute', 'sudo su hdfs - -s /bin/bash -c "export PATH=$PATH:/usr/bin ; yes Y | hdfs --config /etc/hadoop/conf namenode -format"',
|
|
|
|
- path = ['/usr/sbin:/sbin:/usr/local/bin:/bin:/usr/bin'],
|
|
|
|
- )
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', 'yes Y | hdfs --config /etc/hadoop/conf namenode -format',
|
|
|
|
+ path = ['/usr/bin'],
|
|
|
|
+ user = 'hdfs',
|
|
|
|
+ )
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
self.assertResourceCalled('Directory', '/hadoop/hdfs/namenode/namenode-formatted/',
|
|
- recursive = True,
|
|
|
|
- )
|
|
|
|
-
|
|
|
|
|
|
+ recursive = True,
|
|
|
|
+ )
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
self.assertResourceCalled('File', '/etc/hadoop/conf/dfs.exclude',
|
|
owner = 'hdfs',
|
|
owner = 'hdfs',
|
|
content = Template('exclude_hosts_list.j2'),
|
|
content = Template('exclude_hosts_list.j2'),
|
|
@@ -472,14 +470,14 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
path = ['/usr/bin'],
|
|
path = ['/usr/bin'],
|
|
tries = 40,
|
|
tries = 40,
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
user = 'hdfs',
|
|
user = 'hdfs',
|
|
try_sleep = 10,
|
|
try_sleep = 10,
|
|
)
|
|
)
|
|
@@ -513,7 +511,7 @@ class TestNamenode(RMFTestCase):
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
action = ['create'],
|
|
action = ['create'],
|
|
bin_dir = '/usr/bin',
|
|
bin_dir = '/usr/bin',
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn1 | grep active'",
|
|
)
|
|
)
|
|
self.assertNoMoreResources()
|
|
self.assertNoMoreResources()
|
|
|
|
|
|
@@ -558,14 +556,14 @@ class TestNamenode(RMFTestCase):
|
|
action = ['delete'],
|
|
action = ['delete'],
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if='ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c '[RMF_EXPORT_PLACEHOLDER]ulimit -c unlimited ; /usr/lib/hadoop/sbin/hadoop-daemon.sh --config /etc/hadoop/conf start namenode'",
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
environment = {'HADOOP_LIBEXEC_DIR': '/usr/lib/hadoop/libexec'},
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
not_if = 'ls /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid >/dev/null 2>&1 && ps -p `cat /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid` >/dev/null 2>&1',
|
|
)
|
|
)
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
self.assertResourceCalled('Execute', "hadoop dfsadmin -safemode get | grep 'Safe mode is OFF'",
|
|
path = ['/usr/bin'],
|
|
path = ['/usr/bin'],
|
|
tries = 40,
|
|
tries = 40,
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
|
|
user = 'hdfs',
|
|
user = 'hdfs',
|
|
try_sleep = 10,
|
|
try_sleep = 10,
|
|
)
|
|
)
|
|
@@ -599,7 +597,7 @@ class TestNamenode(RMFTestCase):
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
kinit_path_local = '/usr/bin/kinit',
|
|
action = ['create'],
|
|
action = ['create'],
|
|
bin_dir = '/usr/bin',
|
|
bin_dir = '/usr/bin',
|
|
- only_if = "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
|
|
|
|
|
|
+ only_if = "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf haadmin -getServiceState nn2 | grep active'",
|
|
)
|
|
)
|
|
self.assertNoMoreResources()
|
|
self.assertNoMoreResources()
|
|
|
|
|
|
@@ -773,7 +771,7 @@ class TestNamenode(RMFTestCase):
|
|
hdp_stack_version = self.STACK_VERSION,
|
|
hdp_stack_version = self.STACK_VERSION,
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
target = RMFTestCase.TARGET_COMMON_SERVICES
|
|
)
|
|
)
|
|
- self.assertResourceCalled('Execute', "/usr/bin/sudo su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf balancer -threshold -1'",
|
|
|
|
|
|
+ self.assertResourceCalled('Execute', "ambari-sudo.sh su hdfs -l -s /bin/bash -c 'export PATH=/bin:/usr/bin ; hdfs --config /etc/hadoop/conf balancer -threshold -1'",
|
|
logoutput = False,
|
|
logoutput = False,
|
|
on_new_line = FunctionMock('handle_new_line'),
|
|
on_new_line = FunctionMock('handle_new_line'),
|
|
)
|
|
)
|