Browse Source

HADOOP-12731. Remove useless boxing/unboxing code. Contributed by Kousuke Saruta.

Akira Ajisaka 9 years ago
parent
commit
736eb17a79

+ 1 - 1
hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/util/TestZKSignerSecretProvider.java

@@ -35,7 +35,7 @@ public class TestZKSignerSecretProvider {
 
 
   // rollover every 2 sec
   // rollover every 2 sec
   private final int timeout = 4000;
   private final int timeout = 4000;
-  private final long rolloverFrequency = Long.valueOf(timeout / 2);
+  private final long rolloverFrequency = timeout / 2;
 
 
   @Before
   @Before
   public void setup() throws Exception {
   public void setup() throws Exception {

+ 3 - 0
hadoop-common-project/hadoop-common/CHANGES.txt

@@ -1645,6 +1645,9 @@ Release 2.8.0 - UNRELEASED
     HADOOP-12730. Hadoop streaming -mapper and -reducer options are wrongly
     HADOOP-12730. Hadoop streaming -mapper and -reducer options are wrongly
     documented as required. (Kengo Seki via aajisaka)
     documented as required. (Kengo Seki via aajisaka)
 
 
+    HADOOP-12731. Remove useless boxing/unboxing code.
+    (Kousuke Saruta via aajisaka)
+
 Release 2.7.3 - UNRELEASED
 Release 2.7.3 - UNRELEASED
 
 
   INCOMPATIBLE CHANGES
   INCOMPATIBLE CHANGES

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileContext.java

@@ -761,7 +761,7 @@ public class FileContext {
       @Override
       @Override
       public Boolean next(final AbstractFileSystem fs, final Path p) 
       public Boolean next(final AbstractFileSystem fs, final Path p) 
         throws IOException, UnresolvedLinkException {
         throws IOException, UnresolvedLinkException {
-        return Boolean.valueOf(fs.delete(p, recursive));
+        return fs.delete(p, recursive);
       }
       }
     }.resolve(this, absF);
     }.resolve(this, absF);
   }
   }
@@ -895,7 +895,7 @@ public class FileContext {
       @Override
       @Override
       public Boolean next(final AbstractFileSystem fs, final Path p) 
       public Boolean next(final AbstractFileSystem fs, final Path p) 
         throws IOException, UnresolvedLinkException {
         throws IOException, UnresolvedLinkException {
-        return Boolean.valueOf(fs.setReplication(p, replication));
+        return fs.setReplication(p, replication);
       }
       }
     }.resolve(this, absF);
     }.resolve(this, absF);
   }
   }

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ha/SshFenceByTcpPort.java

@@ -256,10 +256,10 @@ public class SshFenceByTcpPort extends Configured
       }
       }
     }
     }
 
 
-    private Integer parseConfiggedPort(String portStr)
+    private int parseConfiggedPort(String portStr)
         throws BadFencingConfigurationException {
         throws BadFencingConfigurationException {
       try {
       try {
-        return Integer.valueOf(portStr);
+        return Integer.parseInt(portStr);
       } catch (NumberFormatException nfe) {
       } catch (NumberFormatException nfe) {
         throw new BadFencingConfigurationException(
         throw new BadFencingConfigurationException(
             "Port number '" + portStr + "' invalid");
             "Port number '" + portStr + "' invalid");

+ 19 - 35
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/AbstractMapWritable.java

@@ -116,55 +116,39 @@ public abstract class AbstractMapWritable implements Writable, Configurable {
         DataInputBuffer in = new DataInputBuffer();
         DataInputBuffer in = new DataInputBuffer();
         in.reset(out.getData(), out.getLength());
         in.reset(out.getData(), out.getLength());
         readFields(in);
         readFields(in);
-        
+
       } catch (IOException e) {
       } catch (IOException e) {
         throw new IllegalArgumentException("map cannot be copied: " +
         throw new IllegalArgumentException("map cannot be copied: " +
             e.getMessage());
             e.getMessage());
       }
       }
-      
+
     } else {
     } else {
       throw new IllegalArgumentException("source map cannot be null");
       throw new IllegalArgumentException("source map cannot be null");
     }
     }
   }
   }
-  
+
   /** constructor. */
   /** constructor. */
   protected AbstractMapWritable() {
   protected AbstractMapWritable() {
     this.conf = new AtomicReference<Configuration>();
     this.conf = new AtomicReference<Configuration>();
 
 
-    addToMap(ArrayWritable.class,
-        Byte.valueOf(Integer.valueOf(-127).byteValue())); 
-    addToMap(BooleanWritable.class,
-        Byte.valueOf(Integer.valueOf(-126).byteValue()));
-    addToMap(BytesWritable.class,
-        Byte.valueOf(Integer.valueOf(-125).byteValue()));
-    addToMap(FloatWritable.class,
-        Byte.valueOf(Integer.valueOf(-124).byteValue()));
-    addToMap(IntWritable.class,
-        Byte.valueOf(Integer.valueOf(-123).byteValue()));
-    addToMap(LongWritable.class,
-        Byte.valueOf(Integer.valueOf(-122).byteValue()));
-    addToMap(MapWritable.class,
-        Byte.valueOf(Integer.valueOf(-121).byteValue()));
-    addToMap(MD5Hash.class,
-        Byte.valueOf(Integer.valueOf(-120).byteValue()));
-    addToMap(NullWritable.class,
-        Byte.valueOf(Integer.valueOf(-119).byteValue()));
-    addToMap(ObjectWritable.class,
-        Byte.valueOf(Integer.valueOf(-118).byteValue()));
-    addToMap(SortedMapWritable.class,
-        Byte.valueOf(Integer.valueOf(-117).byteValue()));
-    addToMap(Text.class,
-        Byte.valueOf(Integer.valueOf(-116).byteValue()));
-    addToMap(TwoDArrayWritable.class,
-        Byte.valueOf(Integer.valueOf(-115).byteValue()));
-    
-    // UTF8 is deprecated so we don't support it
+    addToMap(ArrayWritable.class, (byte)-127);
+    addToMap(BooleanWritable.class, (byte)-126);
+    addToMap(BytesWritable.class, (byte)-125);
+    addToMap(FloatWritable.class, (byte)-124);
+    addToMap(IntWritable.class, (byte)-123);
+    addToMap(LongWritable.class, (byte)-122);
+    addToMap(MapWritable.class, (byte)-121);
+    addToMap(MD5Hash.class, (byte)-120);
+    addToMap(NullWritable.class, (byte)-119);
+    addToMap(ObjectWritable.class, (byte)-118);
+    addToMap(SortedMapWritable.class, (byte)-117);
+    addToMap(Text.class, (byte)-116);
+    addToMap(TwoDArrayWritable.class, (byte)-115);
 
 
-    addToMap(VIntWritable.class,
-        Byte.valueOf(Integer.valueOf(-114).byteValue()));
-    addToMap(VLongWritable.class,
-        Byte.valueOf(Integer.valueOf(-113).byteValue()));
+    // UTF8 is deprecated so we don't support it
 
 
+    addToMap(VIntWritable.class, (byte)-114);
+    addToMap(VLongWritable.class, (byte)-113);
   }
   }
 
 
   /** @return the conf */
   /** @return the conf */

+ 4 - 5
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ShellBasedIdMapping.java

@@ -203,9 +203,8 @@ public class ShellBasedIdMapping implements IdMappingServiceProvider {
    * Integer, e.g. 4294967294 maps to -2 and 4294967295 maps to -1.
    * Integer, e.g. 4294967294 maps to -2 and 4294967295 maps to -1.
    */
    */
   private static Integer parseId(final String idStr) {
   private static Integer parseId(final String idStr) {
-    Long longVal = Long.parseLong(idStr);
-    int intVal = longVal.intValue();
-    return Integer.valueOf(intVal);
+    long longVal = Long.parseLong(idStr);
+    return Integer.valueOf((int)longVal);
   }
   }
   
   
   /**
   /**
@@ -605,8 +604,8 @@ public class ShellBasedIdMapping implements IdMappingServiceProvider {
         // We know the line is fine to parse without error checking like this
         // We know the line is fine to parse without error checking like this
         // since it matched the regex above.
         // since it matched the regex above.
         String firstComponent = lineMatcher.group(1);
         String firstComponent = lineMatcher.group(1);
-        int remoteId = parseId(lineMatcher.group(2));
-        int localId = parseId(lineMatcher.group(3));
+        Integer remoteId = parseId(lineMatcher.group(2));
+        Integer localId = parseId(lineMatcher.group(3));
         if (firstComponent.equals("uid")) {
         if (firstComponent.equals("uid")) {
           uidMapping.put(localId, remoteId);
           uidMapping.put(localId, remoteId);
         } else {
         } else {

+ 2 - 2
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/bloom/Key.java

@@ -177,8 +177,8 @@ public class Key implements WritableComparable<Key> {
     }
     }
     
     
     if (result == 0) {
     if (result == 0) {
-      result = Double.valueOf(this.weight - other.weight).intValue();
+      result = (int)(this.weight - other.weight);
     }
     }
     return result;
     return result;
   }
   }
-}
+}

+ 4 - 4
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ha/TestSshFenceByTcpPort.java

@@ -34,20 +34,20 @@ public class TestSshFenceByTcpPort {
   static {
   static {
     ((Log4JLogger)SshFenceByTcpPort.LOG).getLogger().setLevel(Level.ALL);
     ((Log4JLogger)SshFenceByTcpPort.LOG).getLogger().setLevel(Level.ALL);
   }
   }
-  
+
   private static String TEST_FENCING_HOST = System.getProperty(
   private static String TEST_FENCING_HOST = System.getProperty(
       "test.TestSshFenceByTcpPort.host", "localhost");
       "test.TestSshFenceByTcpPort.host", "localhost");
   private static final String TEST_FENCING_PORT = System.getProperty(
   private static final String TEST_FENCING_PORT = System.getProperty(
       "test.TestSshFenceByTcpPort.port", "8020");
       "test.TestSshFenceByTcpPort.port", "8020");
   private static final String TEST_KEYFILE = System.getProperty(
   private static final String TEST_KEYFILE = System.getProperty(
       "test.TestSshFenceByTcpPort.key");
       "test.TestSshFenceByTcpPort.key");
-  
+
   private static final InetSocketAddress TEST_ADDR =
   private static final InetSocketAddress TEST_ADDR =
     new InetSocketAddress(TEST_FENCING_HOST,
     new InetSocketAddress(TEST_FENCING_HOST,
-      Integer.valueOf(TEST_FENCING_PORT));
+      Integer.parseInt(TEST_FENCING_PORT));
   private static final HAServiceTarget TEST_TARGET =
   private static final HAServiceTarget TEST_TARGET =
     new DummyHAService(HAServiceState.ACTIVE, TEST_ADDR);
     new DummyHAService(HAServiceState.ACTIVE, TEST_ADDR);
-  
+
   /**
   /**
    *  Connect to Google's DNS server - not running ssh!
    *  Connect to Google's DNS server - not running ssh!
    */
    */

+ 1 - 1
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/test/GenericTestUtils.java

@@ -479,6 +479,6 @@ public abstract class GenericTestUtils {
    */
    */
   public static void assumeInNativeProfile() {
   public static void assumeInNativeProfile() {
     Assume.assumeTrue(
     Assume.assumeTrue(
-        Boolean.valueOf(System.getProperty("runningWithNative", "false")));
+        Boolean.parseBoolean(System.getProperty("runningWithNative", "false")));
   }
   }
 }
 }

+ 2 - 2
hadoop-tools/hadoop-gridmix/src/main/java/org/apache/hadoop/mapred/gridmix/DistributedCacheEmulator.java

@@ -330,7 +330,7 @@ class DistributedCacheEmulator {
         // Check if visibilities are available because older hadoop versions
         // Check if visibilities are available because older hadoop versions
         // didn't have public, private Distributed Caches separately.
         // didn't have public, private Distributed Caches separately.
         boolean visibility =
         boolean visibility =
-            (visibilities == null) ? true : Boolean.valueOf(visibilities[i]);
+            (visibilities == null) || Boolean.parseBoolean(visibilities[i]);
         if (isLocalDistCacheFile(files[i], user, visibility)) {
         if (isLocalDistCacheFile(files[i], user, visibility)) {
           // local FS based distributed cache file.
           // local FS based distributed cache file.
           // Create this file on the pseudo local FS on the fly (i.e. when the
           // Create this file on the pseudo local FS on the fly (i.e. when the
@@ -514,7 +514,7 @@ class DistributedCacheEmulator {
           // Check if visibilities are available because older hadoop versions
           // Check if visibilities are available because older hadoop versions
           // didn't have public, private Distributed Caches separately.
           // didn't have public, private Distributed Caches separately.
           boolean visibility =
           boolean visibility =
-            (visibilities == null) ? true : Boolean.valueOf(visibilities[i]);
+              (visibilities == null) || Boolean.parseBoolean(visibilities[i]);
           if (isLocalDistCacheFile(files[i], user, visibility)) {
           if (isLocalDistCacheFile(files[i], user, visibility)) {
             // local FS based distributed cache file.
             // local FS based distributed cache file.
             // Create this file on the pseudo local FS.
             // Create this file on the pseudo local FS.

+ 1 - 1
hadoop-tools/hadoop-gridmix/src/test/java/org/apache/hadoop/mapred/gridmix/CommonJobTest.java

@@ -172,7 +172,7 @@ public class CommonJobTest {
         assertTrue("Gridmix job name is not in the expected format.",
         assertTrue("Gridmix job name is not in the expected format.",
                 jobName.equals(GridmixJob.JOB_NAME_PREFIX + jobSeqNum));
                 jobName.equals(GridmixJob.JOB_NAME_PREFIX + jobSeqNum));
         final FileStatus stat = GridmixTestUtils.dfs.getFileStatus(new Path(
         final FileStatus stat = GridmixTestUtils.dfs.getFileStatus(new Path(
-                GridmixTestUtils.DEST, "" + Integer.valueOf(jobSeqNum)));
+                GridmixTestUtils.DEST, "" + Integer.parseInt(jobSeqNum)));
         assertEquals("Wrong owner for " + jobName, spec.getUser(),
         assertEquals("Wrong owner for " + jobName, spec.getUser(),
                 stat.getOwner());
                 stat.getOwner());
         final int nMaps = spec.getNumberMaps();
         final int nMaps = spec.getNumberMaps();

+ 1 - 1
hadoop-tools/hadoop-gridmix/src/test/java/org/apache/hadoop/mapred/gridmix/TestPseudoLocalFs.java

@@ -95,7 +95,7 @@ public class TestPseudoLocalFs {
 
 
       // validate fileSize
       // validate fileSize
       String[] parts = path.toUri().getPath().split("\\.");
       String[] parts = path.toUri().getPath().split("\\.");
-      long expectedFileSize = Long.valueOf(parts[parts.length - 1]);
+      long expectedFileSize = Long.parseLong(parts[parts.length - 1]);
       assertEquals("Invalid file size.", expectedFileSize, stat.getLen());
       assertEquals("Invalid file size.", expectedFileSize, stat.getLen());
     } else {
     } else {
       assertTrue("getFileStatus() did not throw Exception for invalid file "
       assertTrue("getFileStatus() did not throw Exception for invalid file "

+ 1 - 1
hadoop-tools/hadoop-streaming/src/test/java/org/apache/hadoop/streaming/DelayEchoApp.java

@@ -45,7 +45,7 @@ public class DelayEchoApp {
     int seconds = 5;
     int seconds = 5;
     if (args.length >= 1) {
     if (args.length >= 1) {
       try {
       try {
-        seconds = Integer.valueOf(args[0]);
+        seconds = Integer.parseInt(args[0]);
       } catch (NumberFormatException e) {
       } catch (NumberFormatException e) {
         // just use default 5.
         // just use default 5.
       }
       }