From 56497be34da043dacce40e8e077b67e430e78dbc Mon Sep 17 00:00:00 2001 From: Shawn Pearce Date: Thu, 4 Apr 2013 15:06:57 -0700 Subject: [PATCH] Delete broken DFS read-ahead support This implementation has been proven to deadlock in production server loads. Google has been running with it disabled for a quite a while, as the bugs have been difficult to identify and fix. Instead of suggesting it works and is useful, drop the code. JGit should not advertise support for functionality that is known to be broken. In a few of the places where read-ahead was enabled by DfsReader there is more information about what blocks should be loaded when. During object representation selection, or size lookup, or sending object as-is to a PackWriter, or sending an entire pack as-is the reader knows exactly which blocks are required in the cache, and it also can compute when those will be needed. The broken read-ahead code was stupid and just read a fixed amount ahead of the current offset, which can waste IOs if more precise data was available. DFS systems are usually slow to respond so read-ahead is still a desired feature, but it needs to be rebuilt from scratch and make better use of the offset information. Change-Id: Ibaed8288ec3340cf93eb269dc0f1f23ab5ab1aea --- .../internal/storage/dfs/DfsBlockCache.java | 40 --- .../storage/dfs/DfsBlockCacheConfig.java | 80 ------ .../internal/storage/dfs/DfsPackFile.java | 6 +- .../jgit/internal/storage/dfs/DfsReader.java | 127 ++------- .../ReadAheadRejectedExecutionHandler.java | 61 ----- .../internal/storage/dfs/ReadAheadTask.java | 241 ------------------ .../org/eclipse/jgit/lib/ConfigConstants.java | 6 - 7 files changed, 17 insertions(+), 544 deletions(-) delete mode 100644 org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadRejectedExecutionHandler.java delete mode 100644 org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadTask.java diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCache.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCache.java index 7824193da..a8d797dff 100644 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCache.java +++ b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCache.java @@ -45,13 +45,10 @@ package org.eclipse.jgit.internal.storage.dfs; import java.io.IOException; -import java.util.ArrayList; import java.util.Collection; import java.util.Collections; -import java.util.List; import java.util.Map; import java.util.concurrent.ConcurrentHashMap; -import java.util.concurrent.ThreadPoolExecutor; import java.util.concurrent.atomic.AtomicLong; import java.util.concurrent.atomic.AtomicReferenceArray; import java.util.concurrent.locks.ReentrantLock; @@ -116,8 +113,6 @@ public static void reconfigure(DfsBlockCacheConfig cfg) { cache = nc; if (oc != null) { - if (oc.readAheadService != null) - oc.readAheadService.shutdown(); for (DfsPackFile pack : oc.getPackFiles()) pack.key.cachedSize.set(0); } @@ -153,12 +148,6 @@ public static DfsBlockCache getInstance() { /** As {@link #blockSize} is a power of 2, bits to shift for a / blockSize. */ private final int blockSizeShift; - /** Number of bytes to read-ahead from current read position. */ - private final int readAheadLimit; - - /** Thread pool to handle optimistic read-ahead. */ - private final ThreadPoolExecutor readAheadService; - /** Cache of pack files, indexed by description. */ private final Map packCache; @@ -209,9 +198,6 @@ else if (eb < 4) clockHand = new Ref(new DfsPackKey(), -1, 0, null); clockHand.next = clockHand; - readAheadLimit = cfg.getReadAheadLimit(); - readAheadService = cfg.getReadAheadService(); - packCache = new ConcurrentHashMap( 16, 0.75f, 1); packFiles = Collections.unmodifiableCollection(packCache.values()); @@ -498,32 +484,6 @@ T get(DfsPackKey key, long position) { return val; } - boolean readAhead(ReadableChannel rc, DfsPackKey key, int size, long pos, - long len, DfsReader ctx) { - if (!ctx.wantReadAhead() || readAheadLimit <= 0 || readAheadService == null) - return false; - - int cap = readAheadLimit / size; - long readAheadEnd = pos + readAheadLimit; - List blocks = new ArrayList(cap); - while (pos < readAheadEnd && pos < len) { - long end = Math.min(pos + size, len); - if (!contains(key, pos)) - blocks.add(new ReadAheadTask.BlockFuture(key, pos, end)); - pos = end; - } - if (blocks.isEmpty()) - return false; - - ReadAheadTask task = new ReadAheadTask(this, rc, blocks); - ReadAheadTask.TaskFuture t = new ReadAheadTask.TaskFuture(task); - for (ReadAheadTask.BlockFuture b : blocks) - b.setTask(t); - readAheadService.execute(t); - ctx.startedReadAhead(blocks); - return true; - } - private T scan(HashEntry n, DfsPackKey pack, long position) { Ref r = scanRef(n, pack, position); return r != null ? r.get() : null; diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCacheConfig.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCacheConfig.java index 09ddcdb1a..ca1451a2b 100644 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCacheConfig.java +++ b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsBlockCacheConfig.java @@ -47,15 +47,6 @@ import static org.eclipse.jgit.lib.ConfigConstants.CONFIG_DFS_SECTION; import static org.eclipse.jgit.lib.ConfigConstants.CONFIG_KEY_BLOCK_LIMIT; import static org.eclipse.jgit.lib.ConfigConstants.CONFIG_KEY_BLOCK_SIZE; -import static org.eclipse.jgit.lib.ConfigConstants.CONFIG_KEY_READ_AHEAD_LIMIT; -import static org.eclipse.jgit.lib.ConfigConstants.CONFIG_KEY_READ_AHEAD_THREADS; - -import java.util.concurrent.ArrayBlockingQueue; -import java.util.concurrent.RejectedExecutionHandler; -import java.util.concurrent.ThreadFactory; -import java.util.concurrent.ThreadPoolExecutor; -import java.util.concurrent.TimeUnit; -import java.util.concurrent.atomic.AtomicInteger; import org.eclipse.jgit.lib.Config; @@ -71,10 +62,6 @@ public class DfsBlockCacheConfig { private int blockSize; - private int readAheadLimit; - - private ThreadPoolExecutor readAheadService; - /** Create a default configuration. */ public DfsBlockCacheConfig() { setBlockLimit(32 * MB); @@ -118,40 +105,6 @@ public DfsBlockCacheConfig setBlockSize(final int newSize) { return this; } - /** @return number of bytes to read ahead sequentially by. */ - public int getReadAheadLimit() { - return readAheadLimit; - } - - /** - * @param newSize - * new read-ahead limit, in bytes. - * @return {@code this} - */ - public DfsBlockCacheConfig setReadAheadLimit(final int newSize) { - readAheadLimit = Math.max(0, newSize); - return this; - } - - /** @return service to perform read-ahead of sequential blocks. */ - public ThreadPoolExecutor getReadAheadService() { - return readAheadService; - } - - /** - * @param svc - * service to perform read-ahead of sequential blocks with. If - * not null the {@link RejectedExecutionHandler} must be managed - * by the JGit DFS library and not the application. - * @return {@code this}. - */ - public DfsBlockCacheConfig setReadAheadService(ThreadPoolExecutor svc) { - if (svc != null) - svc.setRejectedExecutionHandler(ReadAheadRejectedExecutionHandler.INSTANCE); - readAheadService = svc; - return this; - } - /** * Update properties by setting fields from the configuration. *

@@ -174,39 +127,6 @@ public DfsBlockCacheConfig fromConfig(final Config rc) { CONFIG_DFS_SECTION, CONFIG_KEY_BLOCK_SIZE, getBlockSize())); - - setReadAheadLimit(rc.getInt( - CONFIG_CORE_SECTION, - CONFIG_DFS_SECTION, - CONFIG_KEY_READ_AHEAD_LIMIT, - getReadAheadLimit())); - - int readAheadThreads = rc.getInt( - CONFIG_CORE_SECTION, - CONFIG_DFS_SECTION, - CONFIG_KEY_READ_AHEAD_THREADS, - 0); - - if (0 < getReadAheadLimit() && 0 < readAheadThreads) { - setReadAheadService(new ThreadPoolExecutor( - 1, // Minimum number of threads kept alive. - readAheadThreads, // Maximum threads active. - 60, TimeUnit.SECONDS, // Idle threads wait this long before ending. - new ArrayBlockingQueue(1), // Do not queue deeply. - new ThreadFactory() { - private final String name = "JGit-DFS-ReadAhead"; //$NON-NLS-1$ - private final AtomicInteger cnt = new AtomicInteger(); - private final ThreadGroup group = new ThreadGroup(name); - - public Thread newThread(Runnable body) { - int id = cnt.incrementAndGet(); - Thread thread = new Thread(group, body, name + "-" + id); //$NON-NLS-1$ - thread.setDaemon(true); - thread.setContextClassLoader(getClass().getClassLoader()); - return thread; - } - }, ReadAheadRejectedExecutionHandler.INSTANCE)); - } return this; } } diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsPackFile.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsPackFile.java index 70d1af2ea..1c588d2c4 100644 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsPackFile.java +++ b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsPackFile.java @@ -680,7 +680,6 @@ DfsBlock readOneBlock(long pos, DfsReader ctx) if (invalid) throw new PackInvalidException(getPackName()); - boolean close = true; ReadableChannel rc = ctx.db.openFile(packDesc, PACK); try { // If the block alignment is not yet known, discover it. Prefer the @@ -739,12 +738,9 @@ else if (size < cache.getBlockSize()) } DfsBlock v = new DfsBlock(key, pos, buf); - if (v.end < len) - close = !cache.readAhead(rc, key, size, v.end, len, ctx); return v; } finally { - if (close) - rc.close(); + rc.close(); } } diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsReader.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsReader.java index ca88b5b85..fff6f301d 100644 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsReader.java +++ b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/DfsReader.java @@ -50,7 +50,6 @@ import static org.eclipse.jgit.lib.Constants.OBJ_TREE; import java.io.IOException; -import java.io.InterruptedIOException; import java.security.MessageDigest; import java.text.MessageFormat; import java.util.ArrayList; @@ -60,10 +59,8 @@ import java.util.Comparator; import java.util.HashSet; import java.util.Iterator; -import java.util.LinkedList; import java.util.List; import java.util.Set; -import java.util.concurrent.ExecutionException; import java.util.zip.DataFormatException; import java.util.zip.Inflater; @@ -92,9 +89,6 @@ import org.eclipse.jgit.lib.ObjectLoader; import org.eclipse.jgit.lib.ObjectReader; import org.eclipse.jgit.lib.ProgressMonitor; -import org.eclipse.jgit.revwalk.ObjectWalk; -import org.eclipse.jgit.revwalk.RevCommit; -import org.eclipse.jgit.revwalk.RevWalk; import org.eclipse.jgit.util.BlockList; /** @@ -118,12 +112,8 @@ public final class DfsReader extends ObjectReader implements ObjectReuseAsIs { private DfsPackFile last; - private boolean wantReadAhead; - private boolean avoidUnreachable; - private List pendingReadAhead; - DfsReader(DfsObjDatabase db) { this.db = db; } @@ -322,8 +312,6 @@ private Iterable> findAll( @Override public AsyncObjectLoaderQueue open( Iterable objectIds, final boolean reportMissing) { - wantReadAhead = true; - Iterable> order; IOException error = null; try { @@ -345,7 +333,6 @@ public boolean next() throws MissingObjectException, IOException { } else if (findAllError != null) { throw findAllError; } else { - cancelReadAhead(); return false; } } @@ -365,12 +352,11 @@ public ObjectLoader open() throws IOException { } public boolean cancel(boolean mayInterruptIfRunning) { - cancelReadAhead(); return true; } public void release() { - cancelReadAhead(); + // Nothing to clean up. } }; } @@ -378,8 +364,6 @@ public void release() { @Override public AsyncObjectSizeQueue getObjectSize( Iterable objectIds, final boolean reportMissing) { - wantReadAhead = true; - Iterable> order; IOException error = null; try { @@ -406,7 +390,6 @@ public boolean next() throws MissingObjectException, IOException { } else if (findAllError != null) { throw findAllError; } else { - cancelReadAhead(); return false; } } @@ -424,31 +407,15 @@ public long getSize() { } public boolean cancel(boolean mayInterruptIfRunning) { - cancelReadAhead(); return true; } public void release() { - cancelReadAhead(); + // Nothing to clean up. } }; } - @Override - public void walkAdviceBeginCommits(RevWalk walk, Collection roots) { - wantReadAhead = true; - } - - @Override - public void walkAdviceBeginTrees(ObjectWalk ow, RevCommit min, RevCommit max) { - wantReadAhead = true; - } - - @Override - public void walkAdviceEnd() { - cancelReadAhead(); - } - @Override public long getObjectSize(AnyObjectId objectId, int typeHint) throws MissingObjectException, IncorrectObjectTypeException, @@ -494,23 +461,18 @@ public void selectObjectRepresentation(PackWriter packer, if (tmp.isEmpty()) continue; Collections.sort(tmp, OFFSET_SORT); - try { - wantReadAhead = true; - PackReverseIndex rev = pack.getReverseIdx(this); - DfsObjectRepresentation rep = new DfsObjectRepresentation( - pack, - packIndex); - for (DfsObjectToPack otp : tmp) { - pack.representation(rep, otp.getOffset(), this, rev); - otp.setOffset(0); - packer.select(otp, rep); - if (!otp.isFound()) { - otp.setFound(); - monitor.update(1); - } + PackReverseIndex rev = pack.getReverseIdx(this); + DfsObjectRepresentation rep = new DfsObjectRepresentation( + pack, + packIndex); + for (DfsObjectToPack otp : tmp) { + pack.representation(rep, otp.getOffset(), this, rev); + otp.setOffset(0); + packer.select(otp, rep); + if (!otp.isFound()) { + otp.setFound(); + monitor.update(1); } - } finally { - cancelReadAhead(); } } } @@ -571,24 +533,13 @@ public void writeObjects(PackOutputStream out, List list) case OBJ_BLOB: Collections.sort(list, WRITE_SORT); } - - try { - wantReadAhead = true; - for (ObjectToPack otp : list) - out.writeObject(otp); - } finally { - cancelReadAhead(); - } + for (ObjectToPack otp : list) + out.writeObject(otp); } public void copyPackAsIs(PackOutputStream out, CachedPack pack, boolean validate) throws IOException { - try { - wantReadAhead = true; - ((DfsCachedPack) pack).copyAsIs(out, validate, this); - } finally { - cancelReadAhead(); - } + ((DfsCachedPack) pack).copyAsIs(out, validate, this); } /** @@ -747,60 +698,14 @@ void pin(DfsPackFile pack, long position) throws IOException { // be cleaned up by the GC during the get for the next window. // So we always clear it, even though we are just going to set // it again. - // block = null; - - if (pendingReadAhead != null) - waitForBlock(pack.key, position); block = pack.getOrLoadBlock(position, this); } } - boolean wantReadAhead() { - return wantReadAhead; - } - - void startedReadAhead(List blocks) { - if (pendingReadAhead == null) - pendingReadAhead = new LinkedList(); - pendingReadAhead.addAll(blocks); - } - - private void cancelReadAhead() { - if (pendingReadAhead != null) { - for (ReadAheadTask.BlockFuture f : pendingReadAhead) - f.cancel(true); - pendingReadAhead = null; - } - wantReadAhead = false; - } - - private void waitForBlock(DfsPackKey key, long position) - throws InterruptedIOException { - Iterator itr = pendingReadAhead.iterator(); - while (itr.hasNext()) { - ReadAheadTask.BlockFuture f = itr.next(); - if (f.contains(key, position)) { - try { - f.get(); - } catch (InterruptedException e) { - throw new InterruptedIOException(); - } catch (ExecutionException e) { - // Exceptions should never be thrown by get(). Ignore - // this and let the normal load paths identify any error. - } - itr.remove(); - if (pendingReadAhead.isEmpty()) - pendingReadAhead = null; - break; - } - } - } - /** Release the current window cursor. */ @Override public void release() { - cancelReadAhead(); last = null; block = null; baseCache = null; diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadRejectedExecutionHandler.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadRejectedExecutionHandler.java deleted file mode 100644 index 1b2c0c413..000000000 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadRejectedExecutionHandler.java +++ /dev/null @@ -1,61 +0,0 @@ -/* - * Copyright (C) 2011, Google Inc. - * and other copyright owners as documented in the project's IP log. - * - * This program and the accompanying materials are made available - * under the terms of the Eclipse Distribution License v1.0 which - * accompanies this distribution, is reproduced below, and is - * available at http://www.eclipse.org/org/documents/edl-v10.php - * - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or - * without modification, are permitted provided that the following - * conditions are met: - * - * - Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * - * - Redistributions in binary form must reproduce the above - * copyright notice, this list of conditions and the following - * disclaimer in the documentation and/or other materials provided - * with the distribution. - * - * - Neither the name of the Eclipse Foundation, Inc. nor the - * names of its contributors may be used to endorse or promote - * products derived from this software without specific prior - * written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND - * CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, - * INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES - * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE - * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR - * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT - * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; - * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER - * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, - * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) - * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF - * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - */ - -package org.eclipse.jgit.internal.storage.dfs; - -import java.util.concurrent.RejectedExecutionHandler; -import java.util.concurrent.ThreadPoolExecutor; - -/** This handler aborts a {@link ReadAheadTask} when the queue is full. */ -final class ReadAheadRejectedExecutionHandler implements - RejectedExecutionHandler { - static final ReadAheadRejectedExecutionHandler INSTANCE = new ReadAheadRejectedExecutionHandler(); - - private ReadAheadRejectedExecutionHandler() { - // Singleton, do not create more instances. - } - - public void rejectedExecution(Runnable r, ThreadPoolExecutor executor) { - ((ReadAheadTask.TaskFuture) r).task.abort(); - } -} diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadTask.java b/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadTask.java deleted file mode 100644 index 8169ec0c4..000000000 --- a/org.eclipse.jgit/src/org/eclipse/jgit/internal/storage/dfs/ReadAheadTask.java +++ /dev/null @@ -1,241 +0,0 @@ -/* - * Copyright (C) 2011, Google Inc. - * and other copyright owners as documented in the project's IP log. - * - * This program and the accompanying materials are made available - * under the terms of the Eclipse Distribution License v1.0 which - * accompanies this distribution, is reproduced below, and is - * available at http://www.eclipse.org/org/documents/edl-v10.php - * - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or - * without modification, are permitted provided that the following - * conditions are met: - * - * - Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * - * - Redistributions in binary form must reproduce the above - * copyright notice, this list of conditions and the following - * disclaimer in the documentation and/or other materials provided - * with the distribution. - * - * - Neither the name of the Eclipse Foundation, Inc. nor the - * names of its contributors may be used to endorse or promote - * products derived from this software without specific prior - * written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND - * CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, - * INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES - * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE - * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR - * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT - * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; - * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER - * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, - * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) - * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF - * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - */ - -package org.eclipse.jgit.internal.storage.dfs; - -import java.io.EOFException; -import java.io.IOException; -import java.util.List; -import java.util.concurrent.Callable; -import java.util.concurrent.CountDownLatch; -import java.util.concurrent.ExecutionException; -import java.util.concurrent.Future; -import java.util.concurrent.TimeUnit; -import java.util.concurrent.TimeoutException; - -import org.eclipse.jgit.util.IO; - -final class ReadAheadTask implements Callable { - private final DfsBlockCache cache; - - private final ReadableChannel channel; - - private final List futures; - - private boolean running; - - ReadAheadTask(DfsBlockCache cache, ReadableChannel channel, - List futures) { - this.cache = cache; - this.channel = channel; - this.futures = futures; - } - - public Void call() { - int idx = 0; - try { - synchronized (this) { - if (channel.isOpen()) - running = true; - else - return null; - } - - long position = channel.position(); - for (; idx < futures.size() && !Thread.interrupted(); idx++) { - BlockFuture f = futures.get(idx); - if (cache.contains(f.pack, f.start)) { - f.done(); - continue; - } - - if (position != f.start) - channel.position(f.start); - - int size = (int) (f.end - f.start); - byte[] buf = new byte[size]; - if (IO.read(channel, buf, 0, size) != size) - throw new EOFException(); - - cache.put(new DfsBlock(f.pack, f.start, buf)); - f.done(); - position = f.end; - } - } catch (IOException err) { - // Ignore read-ahead errors. These will be caught later on. - } finally { - for (; idx < futures.size(); idx++) - futures.get(idx).abort(); - close(); - } - return null; - } - - void abort() { - for (BlockFuture f : futures) - f.abort(); - - synchronized (this) { - if (!running) - close(); - } - } - - private synchronized void close() { - try { - if (channel.isOpen()) - channel.close(); - } catch (IOException err) { - // Ignore close errors on a read-only channel. - } - } - - static final class TaskFuture extends java.util.concurrent.FutureTask { - final ReadAheadTask task; - - TaskFuture(ReadAheadTask task) { - super(task); - this.task = task; - } - - @Override - public boolean cancel(boolean mayInterruptIfRunning) { - if (super.cancel(mayInterruptIfRunning)) { - task.abort(); - return true; - } - return false; - } - } - - /** A scheduled read-ahead block load. */ - static final class BlockFuture implements Future { - private static enum State { - PENDING, DONE, CANCELLED; - } - - private volatile State state; - - private volatile Future task; - - private final CountDownLatch latch; - - final DfsPackKey pack; - - final long start; - - final long end; - - BlockFuture(DfsPackKey key, long start, long end) { - this.state = State.PENDING; - this.latch = new CountDownLatch(1); - this.pack = key; - this.start = start; - this.end = end; - } - - synchronized void setTask(Future task) { - if (state == State.PENDING) - this.task = task; - } - - boolean contains(DfsPackKey want, long pos) { - return pack == want && start <= pos && pos < end; - } - - synchronized void done() { - if (state == State.PENDING) { - latch.countDown(); - state = State.DONE; - task = null; - } - } - - synchronized void abort() { - if (state == State.PENDING) { - latch.countDown(); - state = State.CANCELLED; - task = null; - } - } - - public boolean cancel(boolean mayInterruptIfRunning) { - Future t = task; - if (t == null) - return false; - - boolean r = t.cancel(mayInterruptIfRunning); - abort(); - return r; - } - - public Void get() throws InterruptedException, ExecutionException { - latch.await(); - return null; - } - - public Void get(long timeout, TimeUnit unit) - throws InterruptedException, ExecutionException, - TimeoutException { - if (latch.await(timeout, unit)) - return null; - else - throw new TimeoutException(); - } - - public boolean isCancelled() { - State s = state; - if (s == State.DONE) - return false; - if (s == State.CANCELLED) - return true; - - Future t = task; - return t != null ? t.isCancelled() : true; - } - - public boolean isDone() { - return state == State.DONE; - } - } -} diff --git a/org.eclipse.jgit/src/org/eclipse/jgit/lib/ConfigConstants.java b/org.eclipse.jgit/src/org/eclipse/jgit/lib/ConfigConstants.java index 2c2cef3f5..fc0e06a70 100644 --- a/org.eclipse.jgit/src/org/eclipse/jgit/lib/ConfigConstants.java +++ b/org.eclipse.jgit/src/org/eclipse/jgit/lib/ConfigConstants.java @@ -113,12 +113,6 @@ public class ConfigConstants { /** The "blockSize" key */ public static final String CONFIG_KEY_BLOCK_SIZE = "blockSize"; - /** The "readAheadLimit" key */ - public static final String CONFIG_KEY_READ_AHEAD_LIMIT = "readAheadLimit"; - - /** The "readAheadThreads" key */ - public static final String CONFIG_KEY_READ_AHEAD_THREADS = "readAheadThreads"; - /** The "deltaBaseCacheLimit" key */ public static final String CONFIG_KEY_DELTA_BASE_CACHE_LIMIT = "deltaBaseCacheLimit";