+++ /dev/null
-package org.apache.lucene.facet.search;
-
-import java.io.IOException;
-import java.io.Reader;
-import java.util.HashSet;
-import java.util.Set;
-
-import org.apache.lucene.analysis.Analyzer;
-import org.apache.lucene.analysis.MockAnalyzer;
-import org.apache.lucene.analysis.MockTokenizer;
-import org.apache.lucene.analysis.TokenStream;
-import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
-import org.apache.lucene.analysis.tokenattributes.PayloadAttribute;
-import org.apache.lucene.document.Document;
-import org.apache.lucene.document.Field;
-import org.apache.lucene.index.IndexReader;
-import org.apache.lucene.index.Payload;
-import org.apache.lucene.index.RandomIndexWriter;
-import org.apache.lucene.index.Term;
-import org.apache.lucene.store.Directory;
-import org.junit.Test;
-
-import org.apache.lucene.util.LuceneTestCase;
-import org.apache.lucene.facet.search.CategoryListIterator;
-import org.apache.lucene.facet.search.PayloadIntDecodingIterator;
-import org.apache.lucene.util.UnsafeByteArrayOutputStream;
-import org.apache.lucene.util.encoding.DGapIntEncoder;
-import org.apache.lucene.util.encoding.IntEncoder;
-import org.apache.lucene.util.encoding.SortingIntEncoder;
-import org.apache.lucene.util.encoding.UniqueValuesIntEncoder;
-import org.apache.lucene.util.encoding.VInt8IntEncoder;
-
-/**
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-public class CategoryListIteratorTest extends LuceneTestCase {
-
- private static final class DataTokenStream extends TokenStream {
-
- private int idx;
- private PayloadAttribute payload = addAttribute(PayloadAttribute.class);
- private byte[] buf = new byte[20];
- UnsafeByteArrayOutputStream ubaos = new UnsafeByteArrayOutputStream(buf);
- IntEncoder encoder;
- private boolean exhausted = false;
- private CharTermAttribute term = addAttribute(CharTermAttribute.class);
-
- public DataTokenStream(String text, IntEncoder encoder) throws IOException {
- this.encoder = encoder;
- term.setEmpty().append(text);
- }
-
- public void setIdx(int idx) {
- this.idx = idx;
- exhausted = false;
- }
-
- @Override
- public boolean incrementToken() throws IOException {
- if (exhausted) {
- return false;
- }
-
- int[] values = data[idx];
- ubaos.reInit(buf);
- encoder.reInit(ubaos);
- for (int val : values) {
- encoder.encode(val);
- }
- encoder.close();
- payload.setPayload(new Payload(buf, 0, ubaos.length()));
-
- exhausted = true;
- return true;
- }
-
- }
-
- static final int[][] data = new int[][] {
- new int[] { 1, 2 }, new int[] { 3, 4 }, new int[] { 1, 3 }, new int[] { 1, 2, 3, 4 },
- };
-
- @Test
- public void testPayloadIntDecodingIterator() throws Exception {
- Directory dir = newDirectory();
- DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder(
- new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder()))));
- RandomIndexWriter writer = new RandomIndexWriter(random, dir, newIndexWriterConfig(TEST_VERSION_CURRENT,
- new MockAnalyzer(random, MockTokenizer.KEYWORD, false)).setMergePolicy(newLogMergePolicy()));
- for (int i = 0; i < data.length; i++) {
- dts.setIdx(i);
- Document doc = new Document();
- doc.add(new Field("f", dts));
- writer.addDocument(doc);
- }
- IndexReader reader = writer.getReader();
- writer.close();
-
- CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term(
- "f","1"), dts.encoder.createMatchingDecoder());
- cli.init();
- int totalCategories = 0;
- for (int i = 0; i < data.length; i++) {
- Set<Integer> values = new HashSet<Integer>();
- for (int j = 0; j < data[i].length; j++) {
- values.add(data[i][j]);
- }
- cli.skipTo(i);
- long cat;
- while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) {
- assertTrue("expected category not found: " + cat, values.contains((int) cat));
- totalCategories ++;
- }
- }
- assertEquals("Missing categories!",10,totalCategories);
- reader.close();
- dir.close();
- }
-
- /**
- * Test that a document with no payloads does not confuse the payload decoder.
- */
- @Test
- public void testPayloadIteratorWithInvalidDoc() throws Exception {
- Directory dir = newDirectory();
- DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder(
- new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder()))));
- DataTokenStream dts2 = new DataTokenStream("2",new SortingIntEncoder(
- new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder()))));
- // this test requires that no payloads ever be randomly present!
- final Analyzer noPayloadsAnalyzer = new Analyzer() {
- @Override
- public TokenStream tokenStream(String fieldName, Reader reader) {
- return new MockTokenizer(reader, MockTokenizer.KEYWORD, false);
- }
- };
- // NOTE: test is wired to LogMP... because test relies on certain docids having payloads
- RandomIndexWriter writer = new RandomIndexWriter(random, dir,
- newIndexWriterConfig(TEST_VERSION_CURRENT, noPayloadsAnalyzer).setMergePolicy(newLogMergePolicy()));
- for (int i = 0; i < data.length; i++) {
- dts.setIdx(i);
- Document doc = new Document();
- if (i==0 || i == 2) {
- doc.add(new Field("f", dts)); // only docs 0 & 2 have payloads!
- }
- dts2.setIdx(i);
- doc.add(new Field("f", dts2));
- writer.addDocument(doc);
- writer.commit();
- }
-
- // add more documents to expose the bug.
- // for some reason, this bug is not exposed unless these additional documents are added.
- for (int i = 0; i < 10; ++i) {
- Document d = new Document();
- dts.setIdx(2);
- d.add(new Field("f", dts2));
- writer.addDocument(d);
- if (i %10 == 0) {
- writer.commit();
- }
-
- }
-
- IndexReader reader = writer.getReader();
- writer.close();
-
- CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term(
- "f","1"), dts.encoder.createMatchingDecoder());
- cli.init();
- int totalCats = 0;
- for (int i = 0; i < data.length; i++) {
- // doc no. i
- Set<Integer> values = new HashSet<Integer>();
- for (int j = 0; j < data[i].length; j++) {
- values.add(data[i][j]);
- }
- boolean hasDoc = cli.skipTo(i);
- if (hasDoc) {
- assertTrue("Document "+i+" must not have a payload!", i==0 || i==2 );
- long cat;
- while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) {
- assertTrue("expected category not found: " + cat, values.contains((int) cat));
- ++totalCats;
- }
- } else {
- assertFalse("Document "+i+" must have a payload!", i==0 || i==2 );
- }
-
- }
- assertEquals("Wrong number of total categories!", 4, totalCats);
-
- // Ok.. went through the first 4 docs, now lets try the 6th doc (docid 5)
- assertFalse("Doc #6 (docid=5) should not have a payload!",cli.skipTo(5));
- reader.close();
- dir.close();
- }
-
-}