瀏覽代碼

Revert "HDFS-12914. Addendum patch. Block report leases cause missing blocks until next report. Contributed by Santosh Marella, He Xiaoqiao."

This reverts commit cdc5de6448e429d6cb523b8a61bed8b1cb2fc263.
Wei-Chiu Chuang 5 年之前
父節點
當前提交
a50c35bb81

+ 0 - 156
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestBlockReportLease.java

@@ -1,156 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.hadoop.hdfs.server.blockmanagement;
-
-import org.apache.hadoop.hdfs.HdfsConfiguration;
-import org.apache.hadoop.hdfs.MiniDFSCluster;
-import org.apache.hadoop.hdfs.protocol.BlockListAsLongs;
-import org.apache.hadoop.hdfs.server.datanode.DataNode;
-import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
-import org.apache.hadoop.hdfs.server.protocol.BlockReportContext;
-import org.apache.hadoop.hdfs.server.protocol.DatanodeCommand;
-import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration;
-import org.apache.hadoop.hdfs.server.protocol.DatanodeStorage;
-import org.apache.hadoop.hdfs.server.protocol.FinalizeCommand;
-import org.apache.hadoop.hdfs.server.protocol.HeartbeatResponse;
-import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocols;
-import org.apache.hadoop.hdfs.server.protocol.StorageBlockReport;
-import org.apache.hadoop.hdfs.server.protocol.StorageReport;
-import org.apache.hadoop.test.GenericTestUtils.DelayAnswer;
-import org.junit.Test;
-
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Random;
-import java.util.concurrent.ExecutorService;
-import java.util.concurrent.Executors;
-import java.util.concurrent.Future;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.ArgumentMatchers.any;
-import static org.mockito.Mockito.doAnswer;
-import static org.mockito.Mockito.spy;
-
-/**
- * Tests that BlockReportLease in BlockManager.
- */
-public class TestBlockReportLease {
-
-  /**
-   * Test check lease about one BlockReport with many StorageBlockReport.
-   * Before HDFS-12914, when batch storage report to NameNode, it will check
-   * less for one storage by one, So it could part storage report can
-   * be process normally, however, the rest storage report can not be process
-   * since check lease failed.
-   * After HDFS-12914, NameNode check lease once for every blockreport request,
-   * So this issue will not exist anymore.
-   */
-  @Test
-  public void testCheckBlockReportLease() throws Exception {
-    HdfsConfiguration conf = new HdfsConfiguration();
-    Random rand = new Random();
-
-    try (MiniDFSCluster cluster = new MiniDFSCluster.Builder(conf)
-        .numDataNodes(1).build()) {
-      cluster.waitActive();
-
-      FSNamesystem fsn = cluster.getNamesystem();
-      BlockManager blockManager = fsn.getBlockManager();
-      BlockManager spyBlockManager = spy(blockManager);
-      fsn.setBlockManagerForTesting(spyBlockManager);
-      String poolId = cluster.getNamesystem().getBlockPoolId();
-
-      NamenodeProtocols rpcServer = cluster.getNameNodeRpc();
-
-      // Test based on one DataNode report to Namenode
-      DataNode dn = cluster.getDataNodes().get(0);
-      DatanodeDescriptor datanodeDescriptor = spyBlockManager
-          .getDatanodeManager().getDatanode(dn.getDatanodeId());
-
-      DatanodeRegistration dnRegistration = dn.getDNRegistrationForBP(poolId);
-      StorageReport[] storages = dn.getFSDataset().getStorageReports(poolId);
-
-      // Send heartbeat and request full block report lease
-      HeartbeatResponse hbResponse = rpcServer.sendHeartbeat(
-          dnRegistration, storages, 0, 0, 0, 0, 0, null, true, null, null);
-
-      DelayAnswer delayer = new DelayAnswer(BlockManager.LOG);
-      doAnswer(delayer).when(spyBlockManager).processReport(
-          any(DatanodeStorageInfo.class),
-          any(BlockListAsLongs.class),
-          any(BlockReportContext.class));
-
-      ExecutorService pool = Executors.newFixedThreadPool(1);
-
-      // Trigger sendBlockReport
-      BlockReportContext brContext = new BlockReportContext(1, 0,
-          rand.nextLong(), hbResponse.getFullBlockReportLeaseId(), true);
-      Future<DatanodeCommand> sendBRfuturea = pool.submit(() -> {
-        // Build every storage with 100 blocks for sending report
-        DatanodeStorage[] datanodeStorages
-            = new DatanodeStorage[storages.length];
-        for (int i = 0; i < storages.length; i++) {
-          datanodeStorages[i] = storages[i].getStorage();
-        }
-        StorageBlockReport[] reports = createReports(datanodeStorages, 100);
-
-        // Send blockReport
-        return rpcServer.blockReport(dnRegistration, poolId, reports,
-            brContext);
-      });
-
-      // Wait until BlockManager calls processReport
-      delayer.waitForCall();
-
-      // Remove full block report lease about dn
-      spyBlockManager.getBlockReportLeaseManager()
-          .removeLease(datanodeDescriptor);
-
-      // Allow blockreport to proceed
-      delayer.proceed();
-
-      // Get result, it will not null if process successfully
-      DatanodeCommand datanodeCommand = sendBRfuturea.get();
-      assertTrue(datanodeCommand instanceof FinalizeCommand);
-      assertEquals(poolId, ((FinalizeCommand)datanodeCommand)
-          .getBlockPoolId());
-    }
-  }
-
-  private StorageBlockReport[] createReports(DatanodeStorage[] dnStorages,
-      int numBlocks) {
-    int longsPerBlock = 3;
-    int blockListSize = 2 + numBlocks * longsPerBlock;
-    int numStorages = dnStorages.length;
-    StorageBlockReport[] storageBlockReports
-        = new StorageBlockReport[numStorages];
-    for (int i = 0; i < numStorages; i++) {
-      List<Long> longs = new ArrayList<Long>(blockListSize);
-      longs.add(Long.valueOf(numBlocks));
-      longs.add(0L);
-      for (int j = 0; j < blockListSize; ++j) {
-        longs.add(Long.valueOf(j));
-      }
-      BlockListAsLongs blockList = BlockListAsLongs.decodeLongs(longs);
-      storageBlockReports[i] = new StorageBlockReport(dnStorages[i], blockList);
-    }
-    return storageBlockReports;
-  }
-}