releasenotes.html 34 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654
  1. <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
  2. <html><head>
  3. <meta http-equiv="content-type" content="text/html; charset=UTF-8">
  4. <title>Hadoop 0.18.3 Release Notes</title></head><body>
  5. <font face="sans-serif">
  6. <h1>Hadoop 0.18.3 Release Notes</h1>
  7. Hadoop 0.18.3 fixes serveral problems that may lead to data loss
  8. from the file system. Important changes were made to lease recovery and the management of
  9. block replicas. The bug fixes are listed below.
  10. <ul>
  11. <h2>Changes Since Hadoop 0.18.2</h2>
  12. <ul>
  13. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4823'>HADOOP-4823</a>] - Should not use java.util.NavigableMap in 0.18</li>
  14. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4824'>HADOOP-4824</a>] - Should not use File.setWritable(..) in 0.18</li>
  15. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-1980'>HADOOP-1980</a>] - 'dfsadmin -safemode enter' should prevent the namenode from leaving safemode automatically after startup</li>
  16. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-3121'>HADOOP-3121</a>] - dfs -lsr fail with &quot;Could not get listing &quot;</li>
  17. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-3883'>HADOOP-3883</a>] - TestFileCreation fails once in a while</li>
  18. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4061'>HADOOP-4061</a>] - Large number of decommission freezes the Namenode</li>
  19. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4257'>HADOOP-4257</a>] - TestLeaseRecovery2.testBlockSynchronization failing.</li>
  20. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4499'>HADOOP-4499</a>] - DFSClient should invoke checksumOk only once.</li>
  21. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4542'>HADOOP-4542</a>] - Fault in TestDistributedUpgrade</li>
  22. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4556'>HADOOP-4556</a>] - Block went missing</li>
  23. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4597'>HADOOP-4597</a>] - Under-replicated blocks are not calculated if the name-node is forced out of safe-mode.</li>
  24. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4610'>HADOOP-4610</a>] - Always calculate mis-replicated blocks when safe-mode is turned off.</li>
  25. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4613'>HADOOP-4613</a>] - browseBlock.jsp does not generate &quot;genstamp&quot; property.</li>
  26. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4614'>HADOOP-4614</a>] - &quot;Too many open files&quot; error while processing a large gzip file</li>
  27. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4616'>HADOOP-4616</a>] - assertion makes fuse-dfs exit when reading incomplete data</li>
  28. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4620'>HADOOP-4620</a>] - Streaming mapper never completes if the mapper does not write to stdout</li>
  29. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4635'>HADOOP-4635</a>] - Memory leak ?</li>
  30. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4643'>HADOOP-4643</a>] - NameNode should exclude excessive replicas when counting live replicas for a block</li>
  31. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4647'>HADOOP-4647</a>] - NamenodeFsck creates a new DFSClient but never closes it</li>
  32. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4654'>HADOOP-4654</a>] - remove temporary output directory of failed tasks</li>
  33. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4659'>HADOOP-4659</a>] - Root cause of connection failure is being lost to code that uses it for delaying startup</li>
  34. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4679'>HADOOP-4679</a>] - Datanode prints tons of log messages: Waiting for threadgroup to exit, active theads is XX</li>
  35. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4702'>HADOOP-4702</a>] - Failed block replication leaves an incomplete block in receiver's tmp data directory</li>
  36. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4703'>HADOOP-4703</a>] - DataNode.createInterDataNodeProtocolProxy should not wait for proxy forever while recovering lease</li>
  37. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4713'>HADOOP-4713</a>] - librecordio does not scale to large records</li>
  38. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4714'>HADOOP-4714</a>] - map tasks timing out during merge phase</li>
  39. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4717'>HADOOP-4717</a>] - Removal of default port# in NameNode.getUri() cause a map/reduce job failed to prompt temporay output</li>
  40. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4726'>HADOOP-4726</a>] - documentation typos: &quot;the the&quot;</li>
  41. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4734'>HADOOP-4734</a>] - Some lease recovery codes in 0.19 or trunk should also be committed in 0.18.</li>
  42. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4742'>HADOOP-4742</a>] - Mistake delete replica in hadoop 0.18.1</li>
  43. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4746'>HADOOP-4746</a>] - Job output directory should be normalized</li>
  44. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4778'>HADOOP-4778</a>] - Check for zero size block meta file when updating a block.</li>
  45. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4795'>HADOOP-4795</a>] - Lease monitor may get into an infinite loop</li>
  46. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4797'>HADOOP-4797</a>] - RPC Server can leave a lot of direct buffers </li>
  47. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4806'>HADOOP-4806</a>] - HDFS rename does not work correctly if src contains Java regular expression special characters</li>
  48. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4810'>HADOOP-4810</a>] - Data lost at cluster startup time</li>
  49. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4822'>HADOOP-4822</a>] - 0.18 cannot be compiled in Java 5.</li>
  50. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4840'>HADOOP-4840</a>] - TestNodeCount sometimes fails with NullPointerException</li>
  51. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4904'>HADOOP-4904</a>] - Deadlock while leaving safe mode.</li>
  52. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4910'>HADOOP-4910</a>] - NameNode should exclude corrupt replicas when choosing excessive replicas to delete</li>
  53. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4919'>HADOOP-4919</a>] - [HOD] Provide execute access to JT history directory path for group</li>
  54. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4924'>HADOOP-4924</a>] - Race condition in re-init of TaskTracker</li>
  55. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4935'>HADOOP-4935</a>] - Manual leaving of safe mode may lead to data lost</li>
  56. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4951'>HADOOP-4951</a>] - Lease monitor does not own the LeaseManager lock in changing leases.</li>
  57. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4961'>HADOOP-4961</a>] - ConcurrentModificationException in lease recovery of empty files.</li>
  58. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4971'>HADOOP-4971</a>] - Block report times from datanodes could converge to same time. </li>
  59. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4983'>HADOOP-4983</a>] - Job counters sometimes go down as tasks run without task failures</li>
  60. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4997'>HADOOP-4997</a>] - workaround for tmp file handling on DataNodes in 0.18 (HADOOP-4663)</li>
  61. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5077'>HADOOP-5077</a>] - JavaDoc errors in 0.18.3</li>
  62. <li>[<a href='https://issues.apache.org/jira/browse/HADOOP-3780'>HADOOP-3780</a>] - JobTracker should synchronously resolve the tasktracker's network location when the tracker registers</li>
  63. </ul>
  64. </ul>
  65. <h1>Hadoop 0.18.2 Release Notes</h1>
  66. The bug fixes are listed below.
  67. <ul>
  68. <h2>Changes Since Hadoop 0.18.1</h2>
  69. <ul>
  70. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-2421">HADOOP-2421</a>] - Release JDiff report of changes between different versions of Hadoop.</li>
  71. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3217">HADOOP-3217</a>] - [HOD] Be less aggressive when querying job status from resource manager.</li>
  72. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3614">HADOOP-3614</a>] - TestLeaseRecovery fails when run with assertions enabled.</li>
  73. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3786">HADOOP-3786</a>] - Changes in HOD documentation.</li>
  74. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3914">HADOOP-3914</a>] - checksumOk implementation in DFSClient can break applications.</li>
  75. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4116">HADOOP-4116</a>] - Balancer should provide better resource management.</li>
  76. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4228">HADOOP-4228</a>] - dfs datanode metrics, bytes_read, bytes_written overflows due to incorrect type used.</li>
  77. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4271">HADOOP-4271</a>] - Bug in FSInputChecker makes it possible to read from an invalid buffer.</li>
  78. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4277">HADOOP-4277</a>] - Checksum verification is disabled for LocalFS.</li>
  79. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4292">HADOOP-4292</a>] - append() does not work for LocalFileSystem.</li>
  80. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4314">HADOOP-4314</a>] - TestReplication fails quite often.</li>
  81. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4326">HADOOP-4326</a>] - ChecksumFileSystem does not override all create(...) methods.</li>
  82. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4340">HADOOP-4340</a>] - "hadoop jar" always returns exit code 0 (success) to the shell when jar throws a fatal exception.</li>
  83. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4351">HADOOP-4351</a>] - ArrayIndexOutOfBoundsException during fsck.</li>
  84. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4369">HADOOP-4369</a>] - Metric Averages are not averages.</li>
  85. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4395">HADOOP-4395</a>] - Reloading FSImage and FSEditLog may erase user and group information.</li>
  86. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4399">HADOOP-4399</a>] - fuse-dfs per FD context is not thread safe and can cause segfaults and corruptions.</li>
  87. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4403">HADOOP-4403</a>] - TestLeaseRecovery.testBlockSynchronization failed on trunk.</li>
  88. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4407">HADOOP-4407</a>] - HADOOP-4395 should use a Java 1.5 API for 0.18.</li>
  89. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4467">HADOOP-4467</a>] - SerializationFactory should use current context ClassLoader.</li>
  90. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4469">HADOOP-4469</a>] - ant jar file not being included in tar distribution.</li>
  91. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4483">HADOOP-4483</a>] - getBlockArray in DatanodeDescriptor does not honor passed in maxblocks value.</li>
  92. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4517">HADOOP-4517</a>] - unstable dfs when running jobs on 0.18.1.</li>
  93. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4526">HADOOP-4526</a>] - fsck failing with NullPointerException (return value 0).</li>
  94. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4533">HADOOP-4533</a>] - HDFS client of hadoop 0.18.1 and HDFS server 0.18.2 (0.18 branch) not compatible.</li>
  95. </ul>
  96. </ul>
  97. <h1>Hadoop 0.18.1 Release Notes</h1>
  98. The bug fixes are listed below.
  99. <ul>
  100. <h2>Changes Since Hadoop 0.18.0</h2>
  101. <ul>
  102. <li><a name="changes">[</a><a href="https://issues.apache.org/jira/browse/HADOOP-4040">HADOOP-4040</a>] - Remove the hardcoded ipc.client.connection.maxidletime setting from the TaskTracker.Child.main().</li>
  103. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3934">HADOOP-3934</a>] - Update log4j from 1.2.13 to 1.2.15.</li>
  104. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3995">HADOOP-3995</a>] - renameTo(src, dst) does not restore src name in case of quota failur.</li>
  105. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4046">HADOOP-4046</a>] - WritableComparator's constructor should be protected instead of private.</li>
  106. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3821">HADOOP-3821</a>]
  107. - SequenceFile's Reader.decompressorPool or Writer.decompressorPool
  108. gets into an inconsistent state when calling close() more than onc.</li>
  109. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-3940">HADOOP-3940</a>] - Reduce often attempts in memory merge with no work.</li>
  110. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4161">HADOOP-4161</a>] - [HOD] Uncaught exceptions can potentially hang hod-client.</li>
  111. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4060">HADOOP-4060</a>] - [HOD] Make HOD to roll log files on the client.</li>
  112. <li>[<a href="https://issues.apache.org/jira/browse/HADOOP-4145">HADOOP-4145</a>] - [HOD] Support an accounting plugin script for HOD.</li>
  113. </ul>
  114. </ul>
  115. <h1>Hadoop 0.18.0 Release Notes</h1>
  116. These release notes include new developer and user facing incompatibilities, features, and major improvements.
  117. The table below is sorted by Component.
  118. <ul>
  119. <h2>Changes Since Hadoop 0.17.2</h2>
  120. <ul>
  121. <table 100%="" border="1" cellpadding="4">
  122. <tbody><tr>
  123. <td><b>Issue</b></td>
  124. <td><b>Component</b></td>
  125. <td><b>Notes</b></td>
  126. </tr>
  127. <tr>
  128. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3355">HADOOP-3355</a></td>
  129. <td>conf</td>
  130. <td>Added support for hexadecimal values in
  131. Configuration</td>
  132. </tr>
  133. <tr>
  134. <td><a href="https://issues.apache.org/jira/browse/HADOOP-1702">HADOOP-1702</a></td>
  135. <td>dfs</td>
  136. <td>Reduced buffer copies as data is written to HDFS.
  137. The order of sending data bytes and control information has changed, but this
  138. will not be observed by client applications.</td>
  139. </tr>
  140. <tr>
  141. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2065">HADOOP-2065</a></td>
  142. <td>dfs</td>
  143. <td>Added "corrupt" flag to LocatedBlock to
  144. indicate that all replicas of the block thought to be corrupt.</td>
  145. </tr>
  146. <tr>
  147. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2585">HADOOP-2585</a></td>
  148. <td>dfs</td>
  149. <td>Improved management of replicas of the name space
  150. image. If all replicas on the Name Node are lost, the latest check point can
  151. be loaded from the secondary Name Node. Use parameter
  152. "-importCheckpoint" and specify the location with "fs.checkpoint.dir."
  153. The directory structure on the secondary Name Node has changed to match the
  154. primary Name Node.</td>
  155. </tr>
  156. <tr>
  157. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2656">HADOOP-2656</a></td>
  158. <td>dfs</td>
  159. <td>Associated a generation stamp with each block. On
  160. data nodes, the generation stamp is stored as part of the file name of the
  161. block's meta-data file.</td>
  162. </tr>
  163. <tr>
  164. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2703">HADOOP-2703</a></td>
  165. <td>dfs</td>
  166. <td>Changed fsck to ignore files opened for writing.
  167. Introduced new option "-openforwrite" to explicitly show open
  168. files.</td>
  169. </tr>
  170. <tr>
  171. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2797">HADOOP-2797</a></td>
  172. <td>dfs</td>
  173. <td>Withdrew the upgrade-to-CRC facility. HDFS will no
  174. longer support upgrades from versions without CRCs for block data. Users
  175. upgrading from version 0.13 or earlier must first upgrade to an intermediate
  176. (0.14, 0.15, 0.16, 0.17) version before doing upgrade to version 0.18 or
  177. later.</td>
  178. </tr>
  179. <tr>
  180. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2865">HADOOP-2865</a></td>
  181. <td>dfs</td>
  182. <td>Changed the output of the "fs -ls" command
  183. to more closely match familiar Linux format. Additional changes were made by
  184. HADOOP-3459. Applications that parse the command output should be reviewed.</td>
  185. </tr>
  186. <tr>
  187. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3035">HADOOP-3035</a></td>
  188. <td>dfs</td>
  189. <td>Changed protocol for transferring blocks between
  190. data nodes to report corrupt blocks to data node for re-replication from a
  191. good replica.</td>
  192. </tr>
  193. <tr>
  194. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3113">HADOOP-3113</a></td>
  195. <td>dfs</td>
  196. <td>Added sync() method to FSDataOutputStream to really,
  197. really persist data in HDFS. InterDatanodeProtocol to implement this feature.</td>
  198. </tr>
  199. <tr>
  200. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3164">HADOOP-3164</a></td>
  201. <td>dfs</td>
  202. <td>Changed data node to use FileChannel.tranferTo() to
  203. transfer block data. <br>
  204. </td>
  205. </tr>
  206. <tr>
  207. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3177">HADOOP-3177</a></td>
  208. <td>dfs</td>
  209. <td>Added a new public interface Syncable which declares
  210. the sync() operation. FSDataOutputStream implements Syncable. If the
  211. wrappedStream in FSDataOutputStream is Syncalbe, calling
  212. FSDataOutputStream.sync() is equivalent to call wrappedStream.sync(). Otherwise,
  213. FSDataOutputStream.sync() is a no-op. Both DistributedFileSystem and
  214. LocalFileSystem support the sync() operation.</td>
  215. </tr>
  216. <tr>
  217. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3187">HADOOP-3187</a></td>
  218. <td>dfs</td>
  219. <td>Introduced directory quota as hard limits on the
  220. number of names in the tree rooted at that directory. An administrator may
  221. set quotas on individual directories explicitly. Newly created directories
  222. have no associated quota. File/directory creations fault if the quota would
  223. be exceeded. The attempt to set a quota faults if the directory would be in
  224. violation of the new quota.</td>
  225. </tr>
  226. <tr>
  227. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3193">HADOOP-3193</a></td>
  228. <td>dfs</td>
  229. <td>Added reporter to FSNamesystem stateChangeLog, and a
  230. new metric to track the number of corrupted replicas.</td>
  231. </tr>
  232. <tr>
  233. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3232">HADOOP-3232</a></td>
  234. <td>dfs</td>
  235. <td>Changed 'du' command to run in a seperate thread so
  236. that it does not block user.</td>
  237. </tr>
  238. <tr>
  239. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3310">HADOOP-3310</a></td>
  240. <td>dfs</td>
  241. <td>Implemented Lease Recovery to sync the last bock of
  242. a file. Added ClientDatanodeProtocol for client trigging block recovery.
  243. Changed DatanodeProtocol to support block synchronization. Changed
  244. InterDatanodeProtocol to support block update.</td>
  245. </tr>
  246. <tr>
  247. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3317">HADOOP-3317</a></td>
  248. <td>dfs</td>
  249. <td>Changed the default port for "hdfs:" URIs
  250. to be 8020, so that one may simply use URIs of the form
  251. "hdfs://example.com/dir/file".</td>
  252. </tr>
  253. <tr>
  254. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3329">HADOOP-3329</a></td>
  255. <td>dfs</td>
  256. <td>Changed format of file system image to not store
  257. locations of last block.</td>
  258. </tr>
  259. <tr>
  260. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3336">HADOOP-3336</a></td>
  261. <td>dfs</td>
  262. <td>Added a log4j appender that emits events from
  263. FSNamesystem for audit logging</td>
  264. </tr>
  265. <tr>
  266. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3339">HADOOP-3339</a></td>
  267. <td>dfs</td>
  268. <td>Improved failure handling of last Data Node in write
  269. pipeline. <br>
  270. </td>
  271. </tr>
  272. <tr>
  273. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3390">HADOOP-3390</a></td>
  274. <td>dfs</td>
  275. <td>Removed deprecated
  276. ClientProtocol.abandonFileInProgress().</td>
  277. </tr>
  278. <tr>
  279. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3452">HADOOP-3452</a></td>
  280. <td>dfs</td>
  281. <td>Changed exit status of fsck to report whether the
  282. files system is healthy or corrupt.</td>
  283. </tr>
  284. <tr>
  285. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3459">HADOOP-3459</a></td>
  286. <td>dfs</td>
  287. <td>Changed the output of the "fs -ls" command
  288. to more closely match familiar Linux format. Applications that parse the
  289. command output should be reviewed.</td>
  290. </tr>
  291. <tr>
  292. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3486">HADOOP-3486</a></td>
  293. <td>dfs</td>
  294. <td>Changed the default value of
  295. dfs.blockreport.initialDelay to be 0 seconds.</td>
  296. </tr>
  297. <tr>
  298. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3677">HADOOP-3677</a></td>
  299. <td>dfs</td>
  300. <td>Simplify generation stamp upgrade by making is a
  301. local upgrade on datandodes. Deleted distributed upgrade.</td>
  302. </tr>
  303. <tr>
  304. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2188">HADOOP-2188</a></td>
  305. <td>dfs <br>
  306. ipc</td>
  307. <td>Replaced timeouts with pings to check that client
  308. connection is alive. Removed the property ipc.client.timeout from the default
  309. Hadoop configuration. Removed the metric RpcOpsDiscardedOPsNum. <br>
  310. </td>
  311. </tr>
  312. <tr>
  313. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3283">HADOOP-3283</a></td>
  314. <td>dfs <br>
  315. ipc</td>
  316. <td>Added an IPC server in DataNode and a new IPC
  317. protocol InterDatanodeProtocol. Added conf properties
  318. dfs.datanode.ipc.address and dfs.datanode.handler.count with defaults
  319. "0.0.0.0:50020" and 3, respectively. <br>
  320. Changed the serialization in DatanodeRegistration
  321. and DatanodeInfo, and therefore, updated the versionID in ClientProtocol,
  322. DatanodeProtocol, NamenodeProtocol.</td>
  323. </tr>
  324. <tr>
  325. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3058">HADOOP-3058</a></td>
  326. <td>dfs <br>
  327. metrics</td>
  328. <td>Added FSNamesystem status metrics.</td>
  329. </tr>
  330. <tr>
  331. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3683">HADOOP-3683</a></td>
  332. <td>dfs <br>
  333. metrics</td>
  334. <td>Change FileListed to getNumGetListingOps and add
  335. CreateFileOps, DeleteFileOps and AddBlockOps metrics.</td>
  336. </tr>
  337. <tr>
  338. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3265">HADOOP-3265</a></td>
  339. <td>fs</td>
  340. <td>Removed deprecated API getFileCacheHints</td>
  341. </tr>
  342. <tr>
  343. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3307">HADOOP-3307</a></td>
  344. <td>fs</td>
  345. <td>Introduced archive feature to Hadoop. A Map/Reduce
  346. job can be run to create an archive with indexes. A FileSystem abstraction is
  347. provided over the archive.</td>
  348. </tr>
  349. <tr>
  350. <td><a href="https://issues.apache.org/jira/browse/HADOOP-930">HADOOP-930</a></td>
  351. <td>fs</td>
  352. <td>Added support for reading and writing native S3
  353. files. Native S3 files are referenced using s3n URIs. See
  354. http://wiki.apache.org/hadoop/AmazonS3 for more details.</td>
  355. </tr>
  356. <tr>
  357. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3095">HADOOP-3095</a></td>
  358. <td>fs <br>
  359. fs/s3</td>
  360. <td>Added overloaded method
  361. getFileBlockLocations(FileStatus, long, long). This is an incompatible change
  362. for FileSystem implementations which override getFileBlockLocations(Path,
  363. long, long). They should have the signature of this method changed to
  364. getFileBlockLocations(FileStatus, long, long) to work correctly.</td>
  365. </tr>
  366. <tr>
  367. <td><a href="https://issues.apache.org/jira/browse/HADOOP-4">HADOOP-4</a></td>
  368. <td>fuse-dfs</td>
  369. <td>Introduced FUSE module for HDFS. Module allows mount
  370. of HDFS as a Unix filesystem, and optionally the export of that mount point
  371. to other machines. Writes are disabled. rmdir, mv, mkdir, rm are supported,
  372. but not cp, touch, and the like. Usage information is attached to the Jira
  373. record. <br>
  374. <br>
  375. </td>
  376. </tr>
  377. <tr>
  378. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3184">HADOOP-3184</a></td>
  379. <td>hod</td>
  380. <td>Modified HOD to handle master (NameNode or
  381. JobTracker) failures on bad nodes by trying to bring them up on another node
  382. in the ring. Introduced new property ringmaster.max-master-failures to
  383. specify the maximum number of times a master is allowed to fail.</td>
  384. </tr>
  385. <tr>
  386. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3266">HADOOP-3266</a></td>
  387. <td>hod</td>
  388. <td>Moved HOD change items from CHANGES.txt to a new
  389. file src/contrib/hod/CHANGES.txt.</td>
  390. </tr>
  391. <tr>
  392. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3376">HADOOP-3376</a></td>
  393. <td>hod</td>
  394. <td>Modified HOD client to look for specific messages
  395. related to resource limit overruns and take appropriate actions - such as
  396. either failing to allocate the cluster, or issuing a warning to the user. A
  397. tool is provided, specific to Maui and Torque, that will set these specific
  398. messages.</td>
  399. </tr>
  400. <tr>
  401. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3464">HADOOP-3464</a></td>
  402. <td>hod</td>
  403. <td>Implemented a mechanism to transfer HOD errors that
  404. occur on compute nodes to the submit node running the HOD client, so users
  405. have good feedback on why an allocation failed.</td>
  406. </tr>
  407. <tr>
  408. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3483">HADOOP-3483</a></td>
  409. <td>hod</td>
  410. <td>Modified HOD to create a cluster directory if one
  411. does not exist and to auto-deallocate a cluster while reallocating it, if it
  412. is already dead.</td>
  413. </tr>
  414. <tr>
  415. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3564">HADOOP-3564</a></td>
  416. <td>hod</td>
  417. <td>Modifed HOD to generate the dfs.datanode.ipc.address
  418. parameter in the hadoop-site.xml of datanodes that it launches.</td>
  419. </tr>
  420. <tr>
  421. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3610">HADOOP-3610</a></td>
  422. <td>hod</td>
  423. <td>Modified HOD to automatically create a cluster
  424. directory if the one specified with the script command does not exist.</td>
  425. </tr>
  426. <tr>
  427. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3703">HADOOP-3703</a></td>
  428. <td>hod</td>
  429. <td>Modified logcondense.py to use the new format of
  430. hadoop dfs -lsr output. This version of logcondense would not work with
  431. previous versions of Hadoop and hence is incompatible.</td>
  432. </tr>
  433. <tr>
  434. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3061">HADOOP-3061</a></td>
  435. <td>io</td>
  436. <td>Introduced ByteWritable and DoubleWritable
  437. (implementing WritableComparable) implementations for Byte and Double.</td>
  438. </tr>
  439. <tr>
  440. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3299">HADOOP-3299</a></td>
  441. <td>io <br>
  442. mapred</td>
  443. <td>Changed the TextInputFormat and KeyValueTextInput
  444. classes to initialize the compressionCodecs member variable before
  445. dereferencing it.</td>
  446. </tr>
  447. <tr>
  448. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2909">HADOOP-2909</a></td>
  449. <td>ipc</td>
  450. <td>Removed property ipc.client.maxidletime from the
  451. default configuration. The allowed idle time is twice
  452. ipc.client.connection.maxidletime. <br>
  453. </td>
  454. </tr>
  455. <tr>
  456. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3569">HADOOP-3569</a></td>
  457. <td>KFS</td>
  458. <td>Fixed KFS to have read() read and return 1 byte
  459. instead of 4.</td>
  460. </tr>
  461. <tr>
  462. <td><a href="https://issues.apache.org/jira/browse/HADOOP-1915">HADOOP-1915</a></td>
  463. <td>mapred</td>
  464. <td>Provided a new method to update counters.
  465. "incrCounter(String group, String counter, long amount)"</td>
  466. </tr>
  467. <tr>
  468. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2019">HADOOP-2019</a></td>
  469. <td>mapred</td>
  470. <td>Added support for .tar, .tgz and .tar.gz files in
  471. DistributedCache. File sizes are limited to 2GB.</td>
  472. </tr>
  473. <tr>
  474. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2095">HADOOP-2095</a></td>
  475. <td>mapred</td>
  476. <td>Reduced in-memory copies of keys and values as they
  477. flow through the Map-Reduce framework. Changed the storage of intermediate
  478. map outputs to use new IFile instead of SequenceFile for better compression.</td>
  479. </tr>
  480. <tr>
  481. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2132">HADOOP-2132</a></td>
  482. <td>mapred</td>
  483. <td>Changed "job -kill" to only allow a job
  484. that is in the RUNNING or PREP state to be killed.</td>
  485. </tr>
  486. <tr>
  487. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2181">HADOOP-2181</a></td>
  488. <td>mapred</td>
  489. <td>Added logging for input splits in job tracker log
  490. and job history log. Added web UI for viewing input splits in the job UI and
  491. history UI.</td>
  492. </tr>
  493. <tr>
  494. <td><a href="https://issues.apache.org/jira/browse/HADOOP-236">HADOOP-236</a></td>
  495. <td>mapred</td>
  496. <td>Changed connection protocol job tracker and task
  497. tracker so that task tracker will not connect to a job tracker with a
  498. different build version.</td>
  499. </tr>
  500. <tr>
  501. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2427">HADOOP-2427</a></td>
  502. <td>mapred</td>
  503. <td>The current working directory of a task, i.e.
  504. ${mapred.local.dir}/taskTracker/jobcache/&lt;jobid&gt;/&lt;task_dir&gt;/work
  505. is cleanedup, as soon as the task is finished.</td>
  506. </tr>
  507. <tr>
  508. <td><a href="https://issues.apache.org/jira/browse/HADOOP-2867">HADOOP-2867</a></td>
  509. <td>mapred</td>
  510. <td>Added task's cwd to its LD_LIBRARY_PATH.</td>
  511. </tr>
  512. <tr>
  513. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3135">HADOOP-3135</a></td>
  514. <td>mapred</td>
  515. <td>Changed job submission protocol to not allow
  516. submission if the client's value of mapred.system.dir does not match the job
  517. tracker's. Deprecated JobConf.getSystemDir(); use JobClient.getSystemDir().</td>
  518. </tr>
  519. <tr>
  520. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3221">HADOOP-3221</a></td>
  521. <td>mapred</td>
  522. <td>Added org.apache.hadoop.mapred.lib.NLineInputFormat,
  523. which splits N lines of input as one split. N can be specified by
  524. configuration property "mapred.line.input.format.linespermap",
  525. which defaults to 1.</td>
  526. </tr>
  527. <tr>
  528. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3226">HADOOP-3226</a></td>
  529. <td>mapred</td>
  530. <td>Changed policy for running combiner. The combiner
  531. may be run multiple times as the map's output is sorted and merged.
  532. Additionally, it may be run on the reduce side as data is merged. The old
  533. semantics are available in Hadoop 0.18 if the user calls: <br>
  534. job.setCombineOnlyOnce(true); <br>
  535. </td>
  536. </tr>
  537. <tr>
  538. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3326">HADOOP-3326</a></td>
  539. <td>mapred</td>
  540. <td>Changed fetchOutputs() so that LocalFSMerger and
  541. InMemFSMergeThread threads are spawned only once. The thread gets notified
  542. when something is ready for merge. The merge happens when thresholds are met.</td>
  543. </tr>
  544. <tr>
  545. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3366">HADOOP-3366</a></td>
  546. <td>mapred</td>
  547. <td>Improved shuffle so that all fetched map-outputs are
  548. kept in-memory before being merged by stalling the shuffle so that the
  549. in-memory merge executes and frees up memory for the shuffle.</td>
  550. </tr>
  551. <tr>
  552. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3405">HADOOP-3405</a></td>
  553. <td>mapred</td>
  554. <td>Refactored previously public classes MapTaskStatus,
  555. ReduceTaskStatus, JobSubmissionProtocol, CompletedJobStatusStore to be
  556. package local.</td>
  557. </tr>
  558. <tr>
  559. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3417">HADOOP-3417</a></td>
  560. <td>mapred</td>
  561. <td>Removed the public class
  562. org.apache.hadoop.mapred.JobShell. <br>
  563. Command line options -libjars, -files and -archives are moved to
  564. GenericCommands. Thus applications have to implement
  565. org.apache.hadoop.util.Tool to use the options.</td>
  566. </tr>
  567. <tr>
  568. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3427">HADOOP-3427</a></td>
  569. <td>mapred</td>
  570. <td>Changed shuffle scheduler policy to wait for
  571. notifications from shuffle threads before scheduling more.</td>
  572. </tr>
  573. <tr>
  574. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3460">HADOOP-3460</a></td>
  575. <td>mapred</td>
  576. <td>Created SequenceFileAsBinaryOutputFormat to write
  577. raw bytes as keys and values to a SequenceFile.</td>
  578. </tr>
  579. <tr>
  580. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3512">HADOOP-3512</a></td>
  581. <td>mapred</td>
  582. <td>Separated Distcp, Logalyzer and Archiver into a
  583. tools jar.</td>
  584. </tr>
  585. <tr>
  586. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3565">HADOOP-3565</a></td>
  587. <td>mapred</td>
  588. <td>Change the Java serialization framework, which is
  589. not enabled by default, to correctly make the objects independent of the
  590. previous objects.</td>
  591. </tr>
  592. <tr>
  593. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3598">HADOOP-3598</a></td>
  594. <td>mapred</td>
  595. <td>Changed Map-Reduce framework to no longer create
  596. temporary task output directories for staging outputs if staging outputs
  597. isn't necessary. ${mapred.out.dir}/_temporary/_${taskid}</td>
  598. </tr>
  599. <tr>
  600. <td><a href="https://issues.apache.org/jira/browse/HADOOP-544">HADOOP-544</a></td>
  601. <td>mapred</td>
  602. <td>Introduced new classes JobID, TaskID and
  603. TaskAttemptID, which should be used instead of their string counterparts.
  604. Deprecated functions in JobClient, TaskReport, RunningJob, jobcontrol.Job and
  605. TaskCompletionEvent that use string arguments. Applications can use
  606. xxxID.toString() and xxxID.forName() methods to convert/restore objects
  607. to/from strings. <br>
  608. </td>
  609. </tr>
  610. <tr>
  611. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3230">HADOOP-3230</a></td>
  612. <td>scripts</td>
  613. <td>Added command line tool "job -counter
  614. &lt;job-id&gt; &lt;group-name&gt; &lt;counter-name&gt;" to access
  615. counters.</td>
  616. </tr>
  617. <tr>
  618. <td><a href="https://issues.apache.org/jira/browse/HADOOP-1328">HADOOP-1328</a></td>
  619. <td>streaming</td>
  620. <td>Introduced a way for a streaming process to update
  621. global counters and status using stderr stream to emit information. Use
  622. "reporter:counter:&lt;group&gt;,&lt;counter&gt;,&lt;amount&gt; " to
  623. update a counter. Use "reporter:status:&lt;message&gt;" to update
  624. status.</td>
  625. </tr>
  626. <tr>
  627. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3429">HADOOP-3429</a></td>
  628. <td>streaming</td>
  629. <td>Increased the size of the buffer used in the
  630. communication between the Java task and the Streaming process to 128KB.</td>
  631. </tr>
  632. <tr>
  633. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3379">HADOOP-3379</a></td>
  634. <td>streaming <br>
  635. documentation</td>
  636. <td>Set default value for configuration property
  637. "stream.non.zero.exit.status.is.failure" to be "true".</td>
  638. </tr>
  639. <tr>
  640. <td><a href="https://issues.apache.org/jira/browse/HADOOP-3246">HADOOP-3246</a></td>
  641. <td>util</td>
  642. <td>Introduced an FTPFileSystem backed by Apache Commons
  643. FTPClient to directly store data into HDFS.</td>
  644. </tr>
  645. </tbody></table>
  646. </ul>
  647. </ul>
  648. </font>
  649. </body></html>