capacity-scheduler.xml 7.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210
  1. <!--
  2. Licensed under the Apache License, Version 2.0 (the "License");
  3. you may not use this file except in compliance with the License.
  4. You may obtain a copy of the License at
  5. http://www.apache.org/licenses/LICENSE-2.0
  6. Unless required by applicable law or agreed to in writing, software
  7. distributed under the License is distributed on an "AS IS" BASIS,
  8. WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  9. See the License for the specific language governing permissions and
  10. limitations under the License. See accompanying LICENSE file.
  11. -->
  12. <configuration>
  13. <property>
  14. <name>yarn.scheduler.capacity.maximum-applications</name>
  15. <value>10000</value>
  16. <description>
  17. Maximum number of applications that can be pending and running.
  18. </description>
  19. </property>
  20. <property>
  21. <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
  22. <value>0.1</value>
  23. <description>
  24. Maximum percent of resources in the cluster which can be used to run
  25. application masters i.e. controls number of concurrent running
  26. applications.
  27. </description>
  28. </property>
  29. <property>
  30. <name>yarn.scheduler.capacity.resource-calculator</name>
  31. <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
  32. <description>
  33. The ResourceCalculator implementation to be used to compare
  34. Resources in the scheduler.
  35. The default i.e. DefaultResourceCalculator only uses Memory while
  36. DominantResourceCalculator uses dominant-resource to compare
  37. multi-dimensional resources such as Memory, CPU etc.
  38. </description>
  39. </property>
  40. <property>
  41. <name>yarn.scheduler.capacity.root.queues</name>
  42. <value>default</value>
  43. <description>
  44. The queues at the this level (root is the root queue).
  45. </description>
  46. </property>
  47. <property>
  48. <name>yarn.scheduler.capacity.root.default.capacity</name>
  49. <value>100</value>
  50. <description>Default queue target capacity.</description>
  51. </property>
  52. <property>
  53. <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
  54. <value>1</value>
  55. <description>
  56. Default queue user limit a percentage from 0.0 to 1.0.
  57. </description>
  58. </property>
  59. <property>
  60. <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
  61. <value>100</value>
  62. <description>
  63. The maximum capacity of the default queue.
  64. </description>
  65. </property>
  66. <property>
  67. <name>yarn.scheduler.capacity.root.default.state</name>
  68. <value>RUNNING</value>
  69. <description>
  70. The state of the default queue. State can be one of RUNNING or STOPPED.
  71. </description>
  72. </property>
  73. <property>
  74. <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
  75. <value>*</value>
  76. <description>
  77. The ACL of who can submit jobs to the default queue.
  78. </description>
  79. </property>
  80. <property>
  81. <name>yarn.scheduler.capacity.root.default.acl_administer_queue</name>
  82. <value>*</value>
  83. <description>
  84. The ACL of who can administer jobs on the default queue.
  85. </description>
  86. </property>
  87. <property>
  88. <name>yarn.scheduler.capacity.root.default.acl_application_max_priority</name>
  89. <value>*</value>
  90. <description>
  91. The ACL of who can submit applications with configured priority.
  92. For e.g, [user={name} group={name} max_priority={priority} default_priority={priority}]
  93. </description>
  94. </property>
  95. <property>
  96. <name>yarn.scheduler.capacity.root.default.maximum-application-lifetime
  97. </name>
  98. <value>-1</value>
  99. <description>
  100. Maximum lifetime of an application which is submitted to a queue
  101. in seconds. Any value less than or equal to zero will be considered as
  102. disabled.
  103. This will be a hard time limit for all applications in this
  104. queue. If positive value is configured then any application submitted
  105. to this queue will be killed after exceeds the configured lifetime.
  106. User can also specify lifetime per application basis in
  107. application submission context. But user lifetime will be
  108. overridden if it exceeds queue maximum lifetime. It is point-in-time
  109. configuration.
  110. Note : Configuring too low value will result in killing application
  111. sooner. This feature is applicable only for leaf queue.
  112. </description>
  113. </property>
  114. <property>
  115. <name>yarn.scheduler.capacity.root.default.default-application-lifetime
  116. </name>
  117. <value>-1</value>
  118. <description>
  119. Default lifetime of an application which is submitted to a queue
  120. in seconds. Any value less than or equal to zero will be considered as
  121. disabled.
  122. If the user has not submitted application with lifetime value then this
  123. value will be taken. It is point-in-time configuration.
  124. Note : Default lifetime can't exceed maximum lifetime. This feature is
  125. applicable only for leaf queue.
  126. </description>
  127. </property>
  128. <property>
  129. <name>yarn.scheduler.capacity.node-locality-delay</name>
  130. <value>40</value>
  131. <description>
  132. Number of missed scheduling opportunities after which the CapacityScheduler
  133. attempts to schedule rack-local containers.
  134. When setting this parameter, the size of the cluster should be taken into account.
  135. We use 40 as the default value, which is approximately the number of nodes in one rack.
  136. Note, if this value is -1, the locality constraint in the container request
  137. will be ignored, which disables the delay scheduling.
  138. </description>
  139. </property>
  140. <property>
  141. <name>yarn.scheduler.capacity.rack-locality-additional-delay</name>
  142. <value>-1</value>
  143. <description>
  144. Number of additional missed scheduling opportunities over the node-locality-delay
  145. ones, after which the CapacityScheduler attempts to schedule off-switch containers,
  146. instead of rack-local ones.
  147. Example: with node-locality-delay=40 and rack-locality-delay=20, the scheduler will
  148. attempt rack-local assignments after 40 missed opportunities, and off-switch assignments
  149. after 40+20=60 missed opportunities.
  150. When setting this parameter, the size of the cluster should be taken into account.
  151. We use -1 as the default value, which disables this feature. In this case, the number
  152. of missed opportunities for assigning off-switch containers is calculated based on
  153. the number of containers and unique locations specified in the resource request,
  154. as well as the size of the cluster.
  155. </description>
  156. </property>
  157. <property>
  158. <name>yarn.scheduler.capacity.queue-mappings</name>
  159. <value></value>
  160. <description>
  161. A list of mappings that will be used to assign jobs to queues
  162. The syntax for this list is [u|g]:[name]:[queue_name][,next mapping]*
  163. Typically this list will be used to map users to queues,
  164. for example, u:%user:%user maps all users to queues with the same name
  165. as the user.
  166. </description>
  167. </property>
  168. <property>
  169. <name>yarn.scheduler.capacity.queue-mappings-override.enable</name>
  170. <value>false</value>
  171. <description>
  172. If a queue mapping is present, will it override the value specified
  173. by the user? This can be used by administrators to place jobs in queues
  174. that are different than the one specified by the user.
  175. The default is false.
  176. </description>
  177. </property>
  178. <property>
  179. <name>yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments</name>
  180. <value>1</value>
  181. <description>
  182. Controls the number of OFF_SWITCH assignments allowed
  183. during a node's heartbeat. Increasing this value can improve
  184. scheduling rate for OFF_SWITCH containers. Lower values reduce
  185. "clumping" of applications on particular nodes. The default is 1.
  186. Legal values are 1-MAX_INT. This config is refreshable.
  187. </description>
  188. </property>
  189. </configuration>