|
@@ -0,0 +1,402 @@
|
|
|
+/**
|
|
|
+ * Licensed to the Apache Software Foundation (ASF) under one
|
|
|
+ * or more contributor license agreements. See the NOTICE file
|
|
|
+ * distributed with this work for additional information
|
|
|
+ * regarding copyright ownership. The ASF licenses this file
|
|
|
+ * to you under the Apache License, Version 2.0 (the
|
|
|
+ * "License"); you may not use this file except in compliance
|
|
|
+ * with the License. You may obtain a copy of the License at
|
|
|
+ *
|
|
|
+ * http://www.apache.org/licenses/LICENSE-2.0
|
|
|
+ *
|
|
|
+ * Unless required by applicable law or agreed to in writing, software
|
|
|
+ * distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
+ * See the License for the specific language governing permissions and
|
|
|
+ * limitations under the License.
|
|
|
+ */
|
|
|
+package org.apache.hadoop.hdfs;
|
|
|
+
|
|
|
+import org.apache.commons.logging.impl.Log4JLogger;
|
|
|
+import org.apache.hadoop.conf.Configuration;
|
|
|
+import org.apache.hadoop.fs.BlockLocation;
|
|
|
+import org.apache.hadoop.fs.ChecksumException;
|
|
|
+import org.apache.hadoop.fs.FSDataInputStream;
|
|
|
+import org.apache.hadoop.fs.FSDataOutputStream;
|
|
|
+import org.apache.hadoop.fs.FileSystem;
|
|
|
+import org.apache.hadoop.fs.Path;
|
|
|
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
|
|
|
+import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
|
|
|
+import org.apache.hadoop.io.IOUtils;
|
|
|
+import org.apache.log4j.Level;
|
|
|
+import org.apache.log4j.Logger;
|
|
|
+import org.junit.Before;
|
|
|
+
|
|
|
+import java.io.IOException;
|
|
|
+import java.util.*;
|
|
|
+import java.util.concurrent.atomic.*;
|
|
|
+
|
|
|
+
|
|
|
+/**
|
|
|
+ * This class tests the cases of a concurrent reads/writes to a file;
|
|
|
+ * ie, one writer and one or more readers can see unfinsihed blocks
|
|
|
+ */
|
|
|
+public class TestFileConcurrentReader extends junit.framework.TestCase {
|
|
|
+
|
|
|
+ private enum SyncType
|
|
|
+ {
|
|
|
+ SYNC,
|
|
|
+ APPEND,
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ private static final Logger LOG =
|
|
|
+ Logger.getLogger(TestFileConcurrentReader.class);
|
|
|
+
|
|
|
+ {
|
|
|
+ ((Log4JLogger) LeaseManager.LOG).getLogger().setLevel(Level.ALL);
|
|
|
+ ((Log4JLogger) FSNamesystem.LOG).getLogger().setLevel(Level.ALL);
|
|
|
+ ((Log4JLogger) DFSClient.LOG).getLogger().setLevel(Level.ALL);
|
|
|
+ }
|
|
|
+
|
|
|
+ static final long seed = 0xDEADBEEFL;
|
|
|
+ static final int blockSize = 8192;
|
|
|
+ private static final int DEFAULT_WRITE_SIZE = 1024 + 1;
|
|
|
+ private static final int SMALL_WRITE_SIZE = 61;
|
|
|
+ private Configuration conf;
|
|
|
+ private MiniDFSCluster cluster;
|
|
|
+ private FileSystem fileSystem;
|
|
|
+
|
|
|
+ // creates a file but does not close it
|
|
|
+ private FSDataOutputStream createFile(FileSystem fileSys, Path name, int repl)
|
|
|
+ throws IOException {
|
|
|
+ System.out.println("createFile: Created " + name + " with " + repl + " replica.");
|
|
|
+ FSDataOutputStream stm = fileSys.create(name, true,
|
|
|
+ fileSys.getConf().getInt("io.file.buffer.size", 4096),
|
|
|
+ (short) repl, (long) blockSize);
|
|
|
+ return stm;
|
|
|
+ }
|
|
|
+
|
|
|
+ @Before
|
|
|
+ protected void setUp() throws IOException {
|
|
|
+ conf = new Configuration();
|
|
|
+ init(conf);
|
|
|
+ }
|
|
|
+
|
|
|
+ private void init(Configuration conf) throws IOException {
|
|
|
+ if (cluster != null) {
|
|
|
+ cluster.shutdown();
|
|
|
+ }
|
|
|
+ cluster = new MiniDFSCluster(conf, 1, true, null);
|
|
|
+ cluster.waitClusterUp();
|
|
|
+ fileSystem = cluster.getFileSystem();
|
|
|
+ }
|
|
|
+
|
|
|
+ private void writeFileAndSync(FSDataOutputStream stm, int size)
|
|
|
+ throws IOException {
|
|
|
+// byte[] buffer = AppendTestUtil.randomBytes(seed, size);
|
|
|
+ byte[] buffer = generateSequentialBytes(0, size);
|
|
|
+ stm.write(buffer, 0, size);
|
|
|
+ stm.sync();
|
|
|
+ }
|
|
|
+
|
|
|
+ private void checkCanRead(FileSystem fileSys, Path path, int numBytes)
|
|
|
+ throws IOException {
|
|
|
+ waitForBlocks(fileSys, path);
|
|
|
+ assertBytesAvailable(fileSys, path, numBytes);
|
|
|
+ }
|
|
|
+
|
|
|
+ // make sure bytes are available and match expected
|
|
|
+ private void assertBytesAvailable(
|
|
|
+ FileSystem fileSystem,
|
|
|
+ Path path,
|
|
|
+ int numBytes
|
|
|
+ ) throws IOException {
|
|
|
+ byte[] buffer = new byte[numBytes];
|
|
|
+ FSDataInputStream inputStream = fileSystem.open(path);
|
|
|
+ IOUtils.readFully(inputStream, buffer, 0, numBytes);
|
|
|
+
|
|
|
+ assertTrue(
|
|
|
+ "unable to validate bytes",
|
|
|
+ validateSequentialBytes(buffer, 0, numBytes)
|
|
|
+ );
|
|
|
+ }
|
|
|
+
|
|
|
+ private void waitForBlocks(FileSystem fileSys, Path name)
|
|
|
+ throws IOException {
|
|
|
+ // wait until we have at least one block in the file to read.
|
|
|
+ boolean done = false;
|
|
|
+
|
|
|
+ while (!done) {
|
|
|
+ try {
|
|
|
+ Thread.sleep(1000);
|
|
|
+ } catch (InterruptedException e) {
|
|
|
+ }
|
|
|
+ done = true;
|
|
|
+ BlockLocation[] locations = fileSys.getFileBlockLocations(
|
|
|
+ fileSys.getFileStatus(name), 0, blockSize);
|
|
|
+ if (locations.length < 1) {
|
|
|
+ done = false;
|
|
|
+ continue;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Test that that writes to an incomplete block are available to a reader
|
|
|
+ */
|
|
|
+ public void testUnfinishedBlockRead()
|
|
|
+ throws IOException {
|
|
|
+ try {
|
|
|
+ // check that / exists
|
|
|
+ Path path = new Path("/");
|
|
|
+ System.out.println("Path : \"" + path.toString() + "\"");
|
|
|
+ System.out.println(fileSystem.getFileStatus(path).isDir());
|
|
|
+ assertTrue("/ should be a directory",
|
|
|
+ fileSystem.getFileStatus(path).isDir());
|
|
|
+
|
|
|
+ // create a new file in the root, write data, do no close
|
|
|
+ Path file1 = new Path("/unfinished-block");
|
|
|
+ FSDataOutputStream stm = TestFileCreation.createFile(fileSystem, file1, 1);
|
|
|
+
|
|
|
+ // write partial block and sync
|
|
|
+ int partialBlockSize = blockSize / 2;
|
|
|
+ writeFileAndSync(stm, partialBlockSize);
|
|
|
+
|
|
|
+ // Make sure a client can read it before it is closed
|
|
|
+ checkCanRead(fileSystem, file1, partialBlockSize);
|
|
|
+
|
|
|
+ stm.close();
|
|
|
+ } finally {
|
|
|
+ cluster.shutdown();
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * test case: if the BlockSender decides there is only one packet to send,
|
|
|
+ * the previous computation of the pktSize based on transferToAllowed
|
|
|
+ * would result in too small a buffer to do the buffer-copy needed
|
|
|
+ * for partial chunks.
|
|
|
+ */
|
|
|
+ public void testUnfinishedBlockPacketBufferOverrun() throws IOException {
|
|
|
+ try {
|
|
|
+ // check that / exists
|
|
|
+ Path path = new Path("/");
|
|
|
+ System.out.println("Path : \"" + path.toString() + "\"");
|
|
|
+ System.out.println(fileSystem.getFileStatus(path).isDir());
|
|
|
+ assertTrue("/ should be a directory",
|
|
|
+ fileSystem.getFileStatus(path).isDir());
|
|
|
+
|
|
|
+ // create a new file in the root, write data, do no close
|
|
|
+ Path file1 = new Path("/unfinished-block");
|
|
|
+ final FSDataOutputStream stm =
|
|
|
+ TestFileCreation.createFile(fileSystem, file1, 1);
|
|
|
+
|
|
|
+ // write partial block and sync
|
|
|
+ final int bytesPerChecksum = conf.getInt("io.bytes.per.checksum", 512);
|
|
|
+ final int partialBlockSize = bytesPerChecksum - 1;
|
|
|
+
|
|
|
+ writeFileAndSync(stm, partialBlockSize);
|
|
|
+
|
|
|
+ // Make sure a client can read it before it is closed
|
|
|
+ checkCanRead(fileSystem, file1, partialBlockSize);
|
|
|
+
|
|
|
+ stm.close();
|
|
|
+ } finally {
|
|
|
+ cluster.shutdown();
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ // for some reason, using tranferTo evokes the race condition more often
|
|
|
+ // so test separately
|
|
|
+ public void testUnfinishedBlockCRCErrorTransferTo() throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(true, SyncType.SYNC, DEFAULT_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ public void testUnfinishedBlockCRCErrorTransferToVerySmallWrite()
|
|
|
+ throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(true, SyncType.SYNC, SMALL_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ // fails due to issue w/append, disable
|
|
|
+ public void _testUnfinishedBlockCRCErrorTransferToAppend() throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(true, SyncType.APPEND, DEFAULT_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ public void testUnfinishedBlockCRCErrorNormalTransfer() throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(false, SyncType.SYNC, DEFAULT_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ public void testUnfinishedBlockCRCErrorNormalTransferVerySmallWrite()
|
|
|
+ throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(false, SyncType.SYNC, SMALL_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ // fails due to issue w/append, disable
|
|
|
+ public void _testUnfinishedBlockCRCErrorNormalTransferAppend()
|
|
|
+ throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(false, SyncType.APPEND, DEFAULT_WRITE_SIZE);
|
|
|
+ }
|
|
|
+
|
|
|
+ private void runTestUnfinishedBlockCRCError(
|
|
|
+ final boolean transferToAllowed, SyncType syncType, int writeSize
|
|
|
+ ) throws IOException {
|
|
|
+ runTestUnfinishedBlockCRCError(
|
|
|
+ transferToAllowed, syncType, writeSize, new Configuration()
|
|
|
+ );
|
|
|
+ }
|
|
|
+
|
|
|
+ private void runTestUnfinishedBlockCRCError(
|
|
|
+ final boolean transferToAllowed,
|
|
|
+ final SyncType syncType,
|
|
|
+ final int writeSize,
|
|
|
+ Configuration conf
|
|
|
+ ) throws IOException {
|
|
|
+ try {
|
|
|
+ conf.setBoolean("dfs.support.append", syncType == SyncType.APPEND);
|
|
|
+ conf.setBoolean("dfs.datanode.transferTo.allowed", transferToAllowed);
|
|
|
+ init(conf);
|
|
|
+
|
|
|
+ final Path file = new Path("/block-being-written-to");
|
|
|
+ final int numWrites = 2000;
|
|
|
+ final AtomicBoolean writerDone = new AtomicBoolean(false);
|
|
|
+ final AtomicBoolean writerStarted = new AtomicBoolean(false);
|
|
|
+ final AtomicBoolean error = new AtomicBoolean(false);
|
|
|
+ final FSDataOutputStream initialOutputStream = fileSystem.create(file);
|
|
|
+ Thread writer = new Thread(new Runnable() {
|
|
|
+ private FSDataOutputStream outputStream = initialOutputStream;
|
|
|
+
|
|
|
+ @Override
|
|
|
+ public void run() {
|
|
|
+ try {
|
|
|
+ for (int i = 0; !error.get() && i < numWrites; i++) {
|
|
|
+ try {
|
|
|
+ final byte[] writeBuf =
|
|
|
+ generateSequentialBytes(i * writeSize, writeSize);
|
|
|
+ outputStream.write(writeBuf);
|
|
|
+ if (syncType == SyncType.SYNC) {
|
|
|
+ outputStream.sync();
|
|
|
+ } else { // append
|
|
|
+ outputStream.close();
|
|
|
+ outputStream = fileSystem.append(file);
|
|
|
+ }
|
|
|
+ writerStarted.set(true);
|
|
|
+ } catch (IOException e) {
|
|
|
+ error.set(true);
|
|
|
+ LOG.error(String.format("error writing to file"));
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ outputStream.close();
|
|
|
+ writerDone.set(true);
|
|
|
+ } catch (Exception e) {
|
|
|
+ LOG.error("error in writer", e);
|
|
|
+
|
|
|
+ throw new RuntimeException(e);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ });
|
|
|
+ Thread tailer = new Thread(new Runnable() {
|
|
|
+ @Override
|
|
|
+ public void run() {
|
|
|
+ try {
|
|
|
+ long startPos = 0;
|
|
|
+ while (!writerDone.get() && !error.get()) {
|
|
|
+ if (writerStarted.get()) {
|
|
|
+ try {
|
|
|
+ startPos = tailFile(file, startPos);
|
|
|
+ } catch (IOException e) {
|
|
|
+ LOG.error(String.format("error tailing file %s", file), e);
|
|
|
+
|
|
|
+ throw new RuntimeException(e);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+ } catch (RuntimeException e) {
|
|
|
+ if (e.getCause() instanceof ChecksumException) {
|
|
|
+ error.set(true);
|
|
|
+ }
|
|
|
+
|
|
|
+ LOG.error("error in tailer", e);
|
|
|
+ throw e;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ });
|
|
|
+
|
|
|
+ writer.start();
|
|
|
+ tailer.start();
|
|
|
+
|
|
|
+ try {
|
|
|
+ writer.join();
|
|
|
+ tailer.join();
|
|
|
+
|
|
|
+ assertFalse(
|
|
|
+ "error occurred, see log above", error.get()
|
|
|
+ );
|
|
|
+ } catch (InterruptedException e) {
|
|
|
+ LOG.info("interrupted waiting for writer or tailer to complete");
|
|
|
+
|
|
|
+ Thread.currentThread().interrupt();
|
|
|
+ }
|
|
|
+ initialOutputStream.close();
|
|
|
+ } finally {
|
|
|
+ cluster.shutdown();
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ private byte[] generateSequentialBytes(int start, int length) {
|
|
|
+ byte[] result = new byte[length];
|
|
|
+
|
|
|
+ for (int i = 0; i < length; i++) {
|
|
|
+ result[i] = (byte)((start + i) % 127);
|
|
|
+ }
|
|
|
+
|
|
|
+ return result;
|
|
|
+ }
|
|
|
+
|
|
|
+ private boolean validateSequentialBytes(byte[] buf, int startPos, int len) {
|
|
|
+ for (int i = 0; i < len; i++) {
|
|
|
+ int expected = (i + startPos) % 127;
|
|
|
+
|
|
|
+ if (buf[i] % 127 != expected) {
|
|
|
+ LOG.error(String.format("at position [%d], got [%d] and expected [%d]",
|
|
|
+ startPos, buf[i], expected));
|
|
|
+
|
|
|
+ return false;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return true;
|
|
|
+ }
|
|
|
+
|
|
|
+ private long tailFile(Path file, long startPos) throws IOException {
|
|
|
+ long numRead = 0;
|
|
|
+ FSDataInputStream inputStream = fileSystem.open(file);
|
|
|
+ inputStream.seek(startPos);
|
|
|
+
|
|
|
+ int len = 4 * 1024;
|
|
|
+ byte[] buf = new byte[len];
|
|
|
+ int read;
|
|
|
+ while ((read = inputStream.read(buf)) > -1) {
|
|
|
+ LOG.info(String.format("read %d bytes", read));
|
|
|
+
|
|
|
+ if (!validateSequentialBytes(buf, (int) (startPos + numRead), read)) {
|
|
|
+ LOG.error(String.format("invalid bytes: [%s]\n", Arrays.toString(buf)));
|
|
|
+ throw new ChecksumException(
|
|
|
+ String.format("unable to validate bytes"),
|
|
|
+ startPos
|
|
|
+ );
|
|
|
+ }
|
|
|
+
|
|
|
+ numRead += read;
|
|
|
+ }
|
|
|
+
|
|
|
+ inputStream.close();
|
|
|
+ return numRead + startPos - 1;
|
|
|
+ }
|
|
|
+
|
|
|
+ public void _testClientReadsPastBlockEnd() throws IOException {
|
|
|
+ // todo : client has length greater than server--should handle with
|
|
|
+ // exception, not with corrupt data!!!
|
|
|
+ }
|
|
|
+}
|