--- /dev/null
+package org.apache.lucene.search;
+
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+import java.io.Serializable;
+import java.io.IOException;
+import java.util.Map;
+import java.util.WeakHashMap;
+
+import org.apache.lucene.index.IndexReader;
+import org.apache.lucene.util.FixedBitSet;
+
+/**
+ * Wraps another filter's result and caches it. The purpose is to allow
+ * filters to simply filter, and then wrap with this class
+ * to add caching.
+ *
+ * <p><b>NOTE</b>: if you wrap this filter as a query (eg,
+ * using ConstantScoreQuery), you'll likely want to enforce
+ * deletions (using either {@link DeletesMode#RECACHE} or
+ * {@link DeletesMode#DYNAMIC}).
+ */
+public class CachingWrapperFilter extends Filter {
+ Filter filter;
+
+ /**
+ * Expert: Specifies how new deletions against a reopened
+ * reader should be handled.
+ *
+ * <p>The default is IGNORE, which means the cache entry
+ * will be re-used for a given segment, even when that
+ * segment has been reopened due to changes in deletions.
+ * This is a big performance gain, especially with
+ * near-real-timer readers, since you don't hit a cache
+ * miss on every reopened reader for prior segments.</p>
+ *
+ * <p>However, in some cases this can cause invalid query
+ * results, allowing deleted documents to be returned.
+ * This only happens if the main query does not rule out
+ * deleted documents on its own, such as a toplevel
+ * ConstantScoreQuery. To fix this, use RECACHE to
+ * re-create the cached filter (at a higher per-reopen
+ * cost, but at faster subsequent search performance), or
+ * use DYNAMIC to dynamically intersect deleted docs (fast
+ * reopen time but some hit to search performance).</p>
+ */
+ public static enum DeletesMode {IGNORE, RECACHE, DYNAMIC};
+
+ protected final FilterCache<DocIdSet> cache;
+
+ static abstract class FilterCache<T> implements Serializable {
+
+ /**
+ * A transient Filter cache (package private because of test)
+ */
+ // NOTE: not final so that we can dynamically re-init
+ // after de-serialize
+ transient Map<Object,T> cache;
+
+ private final DeletesMode deletesMode;
+
+ public FilterCache(DeletesMode deletesMode) {
+ this.deletesMode = deletesMode;
+ }
+
+ public synchronized T get(IndexReader reader, Object coreKey, Object delCoreKey) throws IOException {
+ T value;
+
+ if (cache == null) {
+ cache = new WeakHashMap<Object,T>();
+ }
+
+ if (deletesMode == DeletesMode.IGNORE) {
+ // key on core
+ value = cache.get(coreKey);
+ } else if (deletesMode == DeletesMode.RECACHE) {
+ // key on deletes, if any, else core
+ value = cache.get(delCoreKey);
+ } else {
+ assert deletesMode == DeletesMode.DYNAMIC;
+
+ // first try for exact match
+ value = cache.get(delCoreKey);
+
+ if (value == null) {
+ // now for core match, but dynamically AND NOT
+ // deletions
+ value = cache.get(coreKey);
+ if (value != null && reader.hasDeletions()) {
+ value = mergeDeletes(reader, value);
+ }
+ }
+ }
+
+ return value;
+ }
+
+ protected abstract T mergeDeletes(IndexReader reader, T value);
+
+ public synchronized void put(Object coreKey, Object delCoreKey, T value) {
+ if (deletesMode == DeletesMode.IGNORE) {
+ cache.put(coreKey, value);
+ } else if (deletesMode == DeletesMode.RECACHE) {
+ cache.put(delCoreKey, value);
+ } else {
+ cache.put(coreKey, value);
+ cache.put(delCoreKey, value);
+ }
+ }
+ }
+
+ /**
+ * New deletes are ignored by default, which gives higher
+ * cache hit rate on reopened readers. Most of the time
+ * this is safe, because the filter will be AND'd with a
+ * Query that fully enforces deletions. If instead you
+ * need this filter to always enforce deletions, pass
+ * either {@link DeletesMode#RECACHE} or {@link
+ * DeletesMode#DYNAMIC}.
+ * @param filter Filter to cache results of
+ */
+ public CachingWrapperFilter(Filter filter) {
+ this(filter, DeletesMode.IGNORE);
+ }
+
+ /**
+ * Expert: by default, the cached filter will be shared
+ * across reopened segments that only had changes to their
+ * deletions.
+ *
+ * @param filter Filter to cache results of
+ * @param deletesMode See {@link DeletesMode}
+ */
+ public CachingWrapperFilter(Filter filter, DeletesMode deletesMode) {
+ this.filter = filter;
+ cache = new FilterCache<DocIdSet>(deletesMode) {
+ @Override
+ public DocIdSet mergeDeletes(final IndexReader r, final DocIdSet docIdSet) {
+ return new FilteredDocIdSet(docIdSet) {
+ @Override
+ protected boolean match(int docID) {
+ return !r.isDeleted(docID);
+ }
+ };
+ }
+ };
+ }
+
+ /** Provide the DocIdSet to be cached, using the DocIdSet provided
+ * by the wrapped Filter.
+ * <p>This implementation returns the given {@link DocIdSet}, if {@link DocIdSet#isCacheable}
+ * returns <code>true</code>, else it copies the {@link DocIdSetIterator} into
+ * an {@link FixedBitSet}.
+ */
+ protected DocIdSet docIdSetToCache(DocIdSet docIdSet, IndexReader reader) throws IOException {
+ if (docIdSet == null) {
+ // this is better than returning null, as the nonnull result can be cached
+ return DocIdSet.EMPTY_DOCIDSET;
+ } else if (docIdSet.isCacheable()) {
+ return docIdSet;
+ } else {
+ final DocIdSetIterator it = docIdSet.iterator();
+ // null is allowed to be returned by iterator(),
+ // in this case we wrap with the empty set,
+ // which is cacheable.
+ if (it == null) {
+ return DocIdSet.EMPTY_DOCIDSET;
+ } else {
+ final FixedBitSet bits = new FixedBitSet(reader.maxDoc());
+ bits.or(it);
+ return bits;
+ }
+ }
+ }
+
+ // for testing
+ int hitCount, missCount;
+
+ @Override
+ public DocIdSet getDocIdSet(IndexReader reader) throws IOException {
+
+ final Object coreKey = reader.getCoreCacheKey();
+ final Object delCoreKey = reader.hasDeletions() ? reader.getDeletesCacheKey() : coreKey;
+
+ DocIdSet docIdSet = cache.get(reader, coreKey, delCoreKey);
+ if (docIdSet != null) {
+ hitCount++;
+ return docIdSet;
+ }
+
+ missCount++;
+
+ // cache miss
+ docIdSet = docIdSetToCache(filter.getDocIdSet(reader), reader);
+
+ if (docIdSet != null) {
+ cache.put(coreKey, delCoreKey, docIdSet);
+ }
+
+ return docIdSet;
+ }
+
+ @Override
+ public String toString() {
+ return "CachingWrapperFilter("+filter+")";
+ }
+
+ @Override
+ public boolean equals(Object o) {
+ if (!(o instanceof CachingWrapperFilter)) return false;
+ return this.filter.equals(((CachingWrapperFilter)o).filter);
+ }
+
+ @Override
+ public int hashCode() {
+ return filter.hashCode() ^ 0x1117BF25;
+ }
+}