--- /dev/null
+package org.apache.lucene.index;
+
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+import java.util.Iterator;
+import java.util.Map;
+
+import org.apache.lucene.search.Query;
+import org.apache.lucene.util.ArrayUtil;
+import org.apache.lucene.util.RamUsageEstimator;
+import org.apache.lucene.index.BufferedDeletesStream.QueryAndLimit;
+
+/** Holds buffered deletes by term or query, once pushed.
+ * Pushed deletes are write-once, so we shift to more
+ * memory efficient data structure to hold them. We don't
+ * hold docIDs because these are applied on flush. */
+
+class FrozenBufferedDeletes {
+
+ /* Query we often undercount (say 24 bytes), plus int. */
+ final static int BYTES_PER_DEL_QUERY = RamUsageEstimator.NUM_BYTES_OBJECT_REF + RamUsageEstimator.NUM_BYTES_INT + 24;
+
+ // Terms, in sorted order:
+ final PrefixCodedTerms terms;
+ int termCount; // just for debugging
+
+ // Parallel array of deleted query, and the docIDUpto for
+ // each
+ final Query[] queries;
+ final int[] queryLimits;
+ final int bytesUsed;
+ final int numTermDeletes;
+ final long gen;
+
+ public FrozenBufferedDeletes(BufferedDeletes deletes, long gen) {
+ Term termsArray[] = deletes.terms.keySet().toArray(new Term[deletes.terms.size()]);
+ termCount = termsArray.length;
+ ArrayUtil.mergeSort(termsArray);
+ PrefixCodedTerms.Builder builder = new PrefixCodedTerms.Builder();
+ for (Term term : termsArray) {
+ builder.add(term);
+ }
+ terms = builder.finish();
+
+ queries = new Query[deletes.queries.size()];
+ queryLimits = new int[deletes.queries.size()];
+ int upto = 0;
+ for(Map.Entry<Query,Integer> ent : deletes.queries.entrySet()) {
+ queries[upto] = ent.getKey();
+ queryLimits[upto] = ent.getValue();
+ upto++;
+ }
+
+ bytesUsed = (int) terms.getSizeInBytes() + queries.length * BYTES_PER_DEL_QUERY;
+ numTermDeletes = deletes.numTermDeletes.get();
+ this.gen = gen;
+ }
+
+ public Iterable<Term> termsIterable() {
+ return new Iterable<Term>() {
+ // @Override -- not until Java 1.6
+ public Iterator<Term> iterator() {
+ return terms.iterator();
+ }
+ };
+ }
+
+ public Iterable<QueryAndLimit> queriesIterable() {
+ return new Iterable<QueryAndLimit>() {
+ // @Override -- not until Java 1.6
+ public Iterator<QueryAndLimit> iterator() {
+ return new Iterator<QueryAndLimit>() {
+ private int upto;
+
+ // @Override -- not until Java 1.6
+ public boolean hasNext() {
+ return upto < queries.length;
+ }
+
+ // @Override -- not until Java 1.6
+ public QueryAndLimit next() {
+ QueryAndLimit ret = new QueryAndLimit(queries[upto], queryLimits[upto]);
+ upto++;
+ return ret;
+ }
+
+ // @Override -- not until Java 1.6
+ public void remove() {
+ throw new UnsupportedOperationException();
+ }
+ };
+ }
+ };
+ }
+
+ @Override
+ public String toString() {
+ String s = "";
+ if (numTermDeletes != 0) {
+ s += " " + numTermDeletes + " deleted terms (unique count=" + termCount + ")";
+ }
+ if (queries.length != 0) {
+ s += " " + queries.length + " deleted queries";
+ }
+ if (bytesUsed != 0) {
+ s += " bytesUsed=" + bytesUsed;
+ }
+
+ return s;
+ }
+
+ boolean any() {
+ return termCount > 0 || queries.length > 0;
+ }
+}