--- /dev/null
+package org.apache.lucene.facet.search;
+
+import java.io.IOException;
+import java.io.Reader;
+import java.util.HashSet;
+import java.util.Set;
+
+import org.apache.lucene.analysis.Analyzer;
+import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.analysis.MockTokenizer;
+import org.apache.lucene.analysis.TokenStream;
+import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
+import org.apache.lucene.analysis.tokenattributes.PayloadAttribute;
+import org.apache.lucene.document.Document;
+import org.apache.lucene.document.Field;
+import org.apache.lucene.document.Field.Index;
+import org.apache.lucene.document.Field.Store;
+import org.apache.lucene.index.IndexReader;
+import org.apache.lucene.index.Payload;
+import org.apache.lucene.index.RandomIndexWriter;
+import org.apache.lucene.index.Term;
+import org.apache.lucene.store.Directory;
+import org.junit.Test;
+
+import org.apache.lucene.util.LuceneTestCase;
+import org.apache.lucene.facet.search.CategoryListIterator;
+import org.apache.lucene.facet.search.PayloadIntDecodingIterator;
+import org.apache.lucene.util.UnsafeByteArrayOutputStream;
+import org.apache.lucene.util.encoding.DGapIntEncoder;
+import org.apache.lucene.util.encoding.IntEncoder;
+import org.apache.lucene.util.encoding.SortingIntEncoder;
+import org.apache.lucene.util.encoding.UniqueValuesIntEncoder;
+import org.apache.lucene.util.encoding.VInt8IntEncoder;
+
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+public class CategoryListIteratorTest extends LuceneTestCase {
+
+ private static final class DataTokenStream extends TokenStream {
+
+ private int idx;
+ private PayloadAttribute payload = addAttribute(PayloadAttribute.class);
+ private byte[] buf = new byte[20];
+ UnsafeByteArrayOutputStream ubaos = new UnsafeByteArrayOutputStream(buf);
+ IntEncoder encoder;
+ private boolean exhausted = false;
+ private CharTermAttribute term = addAttribute(CharTermAttribute.class);
+
+ public DataTokenStream(String text, IntEncoder encoder) throws IOException {
+ this.encoder = encoder;
+ term.setEmpty().append(text);
+ }
+
+ public void setIdx(int idx) {
+ this.idx = idx;
+ exhausted = false;
+ }
+
+ @Override
+ public boolean incrementToken() throws IOException {
+ if (exhausted) {
+ return false;
+ }
+
+ int[] values = data[idx];
+ ubaos.reInit(buf);
+ encoder.reInit(ubaos);
+ for (int val : values) {
+ encoder.encode(val);
+ }
+ encoder.close();
+ payload.setPayload(new Payload(buf, 0, ubaos.length()));
+
+ exhausted = true;
+ return true;
+ }
+
+ }
+
+ static final int[][] data = new int[][] {
+ new int[] { 1, 2 }, new int[] { 3, 4 }, new int[] { 1, 3 }, new int[] { 1, 2, 3, 4 },
+ };
+
+ @Test
+ public void testPayloadIntDecodingIterator() throws Exception {
+ Directory dir = newDirectory();
+ DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder(
+ new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder()))));
+ RandomIndexWriter writer = new RandomIndexWriter(random, dir, newIndexWriterConfig(TEST_VERSION_CURRENT,
+ new MockAnalyzer(random, MockTokenizer.KEYWORD, false)).setMergePolicy(newLogMergePolicy()));
+ for (int i = 0; i < data.length; i++) {
+ dts.setIdx(i);
+ Document doc = new Document();
+ doc.add(new Field("f", dts));
+ writer.addDocument(doc);
+ }
+ IndexReader reader = writer.getReader();
+ writer.close();
+
+ CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term(
+ "f","1"), dts.encoder.createMatchingDecoder());
+ cli.init();
+ int totalCategories = 0;
+ for (int i = 0; i < data.length; i++) {
+ Set<Integer> values = new HashSet<Integer>();
+ for (int j = 0; j < data[i].length; j++) {
+ values.add(data[i][j]);
+ }
+ cli.skipTo(i);
+ long cat;
+ while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) {
+ assertTrue("expected category not found: " + cat, values.contains((int) cat));
+ totalCategories ++;
+ }
+ }
+ assertEquals("Missing categories!",10,totalCategories);
+ reader.close();
+ dir.close();
+ }
+
+ /**
+ * Test that a document with no payloads does not confuse the payload decoder.
+ */
+ @Test
+ public void testPayloadIteratorWithInvalidDoc() throws Exception {
+ Directory dir = newDirectory();
+ DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder(
+ new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder()))));
+ // this test requires that no payloads ever be randomly present!
+ final Analyzer noPayloadsAnalyzer = new Analyzer() {
+ @Override
+ public TokenStream tokenStream(String fieldName, Reader reader) {
+ return new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
+ }
+ };
+ // NOTE: test is wired to LogMP... because test relies on certain docids having payloads
+ RandomIndexWriter writer = new RandomIndexWriter(random, dir,
+ newIndexWriterConfig(TEST_VERSION_CURRENT, noPayloadsAnalyzer).setMergePolicy(newLogMergePolicy()));
+ for (int i = 0; i < data.length; i++) {
+ Document doc = new Document();
+ if (i == 0) {
+ dts.setIdx(i);
+ doc.add(new Field("f", dts)); // only doc 0 has payloads!
+ } else {
+ doc.add(new Field("f", "1", Store.NO, Index.ANALYZED));
+ }
+ writer.addDocument(doc);
+ writer.commit();
+ }
+
+ IndexReader reader = writer.getReader();
+ writer.close();
+
+ CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term(
+ "f","1"), dts.encoder.createMatchingDecoder());
+ assertTrue("Failed to initialize payload iterator", cli.init());
+ int totalCats = 0;
+ for (int i = 0; i < data.length; i++) {
+ // doc no. i
+ Set<Integer> values = new HashSet<Integer>();
+ for (int j = 0; j < data[i].length; j++) {
+ values.add(data[i][j]);
+ }
+ boolean hasDoc = cli.skipTo(i);
+ if (hasDoc) {
+ assertTrue("Document " + i + " must not have a payload!", i == 0);
+ long cat;
+ while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) {
+ assertTrue("expected category not found: " + cat, values.contains((int) cat));
+ ++totalCats;
+ }
+ } else {
+ assertFalse("Document " + i + " must have a payload!", i == 0);
+ }
+
+ }
+ assertEquals("Wrong number of total categories!", 2, totalCats);
+
+ reader.close();
+ dir.close();
+ }
+
+}