|
@@ -714,6 +714,80 @@
|
|
|
</description>
|
|
|
</property>
|
|
|
|
|
|
+<property>
|
|
|
+ <name>dfs.datanode.readahead.bytes</name>
|
|
|
+ <value>4193404</value>
|
|
|
+ <description>
|
|
|
+ While reading block files, if the Hadoop native libraries are available,
|
|
|
+ the datanode can use the posix_fadvise system call to explicitly
|
|
|
+ page data into the operating system buffer cache ahead of the current
|
|
|
+ reader's position. This can improve performance especially when
|
|
|
+ disks are highly contended.
|
|
|
+
|
|
|
+ This configuration specifies the number of bytes ahead of the current
|
|
|
+ read position which the datanode will attempt to read ahead. This
|
|
|
+ feature may be disabled by configuring this property to 0.
|
|
|
+
|
|
|
+ If the native libraries are not available, this configuration has no
|
|
|
+ effect.
|
|
|
+ </description>
|
|
|
+</property>
|
|
|
+
|
|
|
+<property>
|
|
|
+ <name>dfs.datanode.drop.cache.behind.reads</name>
|
|
|
+ <value>false</value>
|
|
|
+ <description>
|
|
|
+ In some workloads, the data read from HDFS is known to be significantly
|
|
|
+ large enough that it is unlikely to be useful to cache it in the
|
|
|
+ operating system buffer cache. In this case, the DataNode may be
|
|
|
+ configured to automatically purge all data from the buffer cache
|
|
|
+ after it is delivered to the client. This behavior is automatically
|
|
|
+ disabled for workloads which read only short sections of a block
|
|
|
+ (e.g HBase random-IO workloads).
|
|
|
+
|
|
|
+ This may improve performance for some workloads by freeing buffer
|
|
|
+ cache spage usage for more cacheable data.
|
|
|
+
|
|
|
+ If the Hadoop native libraries are not available, this configuration
|
|
|
+ has no effect.
|
|
|
+ </description>
|
|
|
+</property>
|
|
|
+
|
|
|
+<property>
|
|
|
+ <name>dfs.datanode.drop.cache.behind.writes</name>
|
|
|
+ <value>false</value>
|
|
|
+ <description>
|
|
|
+ In some workloads, the data written to HDFS is known to be significantly
|
|
|
+ large enough that it is unlikely to be useful to cache it in the
|
|
|
+ operating system buffer cache. In this case, the DataNode may be
|
|
|
+ configured to automatically purge all data from the buffer cache
|
|
|
+ after it is written to disk.
|
|
|
+
|
|
|
+ This may improve performance for some workloads by freeing buffer
|
|
|
+ cache spage usage for more cacheable data.
|
|
|
+
|
|
|
+ If the Hadoop native libraries are not available, this configuration
|
|
|
+ has no effect.
|
|
|
+ </description>
|
|
|
+</property>
|
|
|
+
|
|
|
+<property>
|
|
|
+ <name>dfs.datanode.sync.behind.writes</name>
|
|
|
+ <value>false</value>
|
|
|
+ <description>
|
|
|
+ If this configuration is enabled, the datanode will instruct the
|
|
|
+ operating system to enqueue all written data to the disk immediately
|
|
|
+ after it is written. This differs from the usual OS policy which
|
|
|
+ may wait for up to 30 seconds before triggering writeback.
|
|
|
+
|
|
|
+ This may improve performance for some workloads by smoothing the
|
|
|
+ IO profile for data written to disk.
|
|
|
+
|
|
|
+ If the Hadoop native libraries are not available, this configuration
|
|
|
+ has no effect.
|
|
|
+ </description>
|
|
|
+</property>
|
|
|
+
|
|
|
<property>
|
|
|
<name>dfs.client.failover.max.attempts</name>
|
|
|
<value>15</value>
|