X-Git-Url: https://git.mdrn.pl/pylucene.git/blobdiff_plain/a2e61f0c04805cfcb8706176758d1283c7e3a55c..aaeed5504b982cf3545252ab528713250aa33eed:/lucene-java-3.4.0/lucene/contrib/facet/src/test/org/apache/lucene/facet/search/CategoryListIteratorTest.java diff --git a/lucene-java-3.4.0/lucene/contrib/facet/src/test/org/apache/lucene/facet/search/CategoryListIteratorTest.java b/lucene-java-3.4.0/lucene/contrib/facet/src/test/org/apache/lucene/facet/search/CategoryListIteratorTest.java deleted file mode 100644 index 1358a70..0000000 --- a/lucene-java-3.4.0/lucene/contrib/facet/src/test/org/apache/lucene/facet/search/CategoryListIteratorTest.java +++ /dev/null @@ -1,213 +0,0 @@ -package org.apache.lucene.facet.search; - -import java.io.IOException; -import java.io.Reader; -import java.util.HashSet; -import java.util.Set; - -import org.apache.lucene.analysis.Analyzer; -import org.apache.lucene.analysis.MockAnalyzer; -import org.apache.lucene.analysis.MockTokenizer; -import org.apache.lucene.analysis.TokenStream; -import org.apache.lucene.analysis.tokenattributes.CharTermAttribute; -import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; -import org.apache.lucene.document.Document; -import org.apache.lucene.document.Field; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.Payload; -import org.apache.lucene.index.RandomIndexWriter; -import org.apache.lucene.index.Term; -import org.apache.lucene.store.Directory; -import org.junit.Test; - -import org.apache.lucene.util.LuceneTestCase; -import org.apache.lucene.facet.search.CategoryListIterator; -import org.apache.lucene.facet.search.PayloadIntDecodingIterator; -import org.apache.lucene.util.UnsafeByteArrayOutputStream; -import org.apache.lucene.util.encoding.DGapIntEncoder; -import org.apache.lucene.util.encoding.IntEncoder; -import org.apache.lucene.util.encoding.SortingIntEncoder; -import org.apache.lucene.util.encoding.UniqueValuesIntEncoder; -import org.apache.lucene.util.encoding.VInt8IntEncoder; - -/** - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the "License"); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -public class CategoryListIteratorTest extends LuceneTestCase { - - private static final class DataTokenStream extends TokenStream { - - private int idx; - private PayloadAttribute payload = addAttribute(PayloadAttribute.class); - private byte[] buf = new byte[20]; - UnsafeByteArrayOutputStream ubaos = new UnsafeByteArrayOutputStream(buf); - IntEncoder encoder; - private boolean exhausted = false; - private CharTermAttribute term = addAttribute(CharTermAttribute.class); - - public DataTokenStream(String text, IntEncoder encoder) throws IOException { - this.encoder = encoder; - term.setEmpty().append(text); - } - - public void setIdx(int idx) { - this.idx = idx; - exhausted = false; - } - - @Override - public boolean incrementToken() throws IOException { - if (exhausted) { - return false; - } - - int[] values = data[idx]; - ubaos.reInit(buf); - encoder.reInit(ubaos); - for (int val : values) { - encoder.encode(val); - } - encoder.close(); - payload.setPayload(new Payload(buf, 0, ubaos.length())); - - exhausted = true; - return true; - } - - } - - static final int[][] data = new int[][] { - new int[] { 1, 2 }, new int[] { 3, 4 }, new int[] { 1, 3 }, new int[] { 1, 2, 3, 4 }, - }; - - @Test - public void testPayloadIntDecodingIterator() throws Exception { - Directory dir = newDirectory(); - DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder( - new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder())))); - RandomIndexWriter writer = new RandomIndexWriter(random, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, - new MockAnalyzer(random, MockTokenizer.KEYWORD, false)).setMergePolicy(newLogMergePolicy())); - for (int i = 0; i < data.length; i++) { - dts.setIdx(i); - Document doc = new Document(); - doc.add(new Field("f", dts)); - writer.addDocument(doc); - } - IndexReader reader = writer.getReader(); - writer.close(); - - CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term( - "f","1"), dts.encoder.createMatchingDecoder()); - cli.init(); - int totalCategories = 0; - for (int i = 0; i < data.length; i++) { - Set values = new HashSet(); - for (int j = 0; j < data[i].length; j++) { - values.add(data[i][j]); - } - cli.skipTo(i); - long cat; - while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) { - assertTrue("expected category not found: " + cat, values.contains((int) cat)); - totalCategories ++; - } - } - assertEquals("Missing categories!",10,totalCategories); - reader.close(); - dir.close(); - } - - /** - * Test that a document with no payloads does not confuse the payload decoder. - */ - @Test - public void testPayloadIteratorWithInvalidDoc() throws Exception { - Directory dir = newDirectory(); - DataTokenStream dts = new DataTokenStream("1",new SortingIntEncoder( - new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder())))); - DataTokenStream dts2 = new DataTokenStream("2",new SortingIntEncoder( - new UniqueValuesIntEncoder(new DGapIntEncoder(new VInt8IntEncoder())))); - // this test requires that no payloads ever be randomly present! - final Analyzer noPayloadsAnalyzer = new Analyzer() { - @Override - public TokenStream tokenStream(String fieldName, Reader reader) { - return new MockTokenizer(reader, MockTokenizer.KEYWORD, false); - } - }; - // NOTE: test is wired to LogMP... because test relies on certain docids having payloads - RandomIndexWriter writer = new RandomIndexWriter(random, dir, - newIndexWriterConfig(TEST_VERSION_CURRENT, noPayloadsAnalyzer).setMergePolicy(newLogMergePolicy())); - for (int i = 0; i < data.length; i++) { - dts.setIdx(i); - Document doc = new Document(); - if (i==0 || i == 2) { - doc.add(new Field("f", dts)); // only docs 0 & 2 have payloads! - } - dts2.setIdx(i); - doc.add(new Field("f", dts2)); - writer.addDocument(doc); - writer.commit(); - } - - // add more documents to expose the bug. - // for some reason, this bug is not exposed unless these additional documents are added. - for (int i = 0; i < 10; ++i) { - Document d = new Document(); - dts.setIdx(2); - d.add(new Field("f", dts2)); - writer.addDocument(d); - if (i %10 == 0) { - writer.commit(); - } - - } - - IndexReader reader = writer.getReader(); - writer.close(); - - CategoryListIterator cli = new PayloadIntDecodingIterator(reader, new Term( - "f","1"), dts.encoder.createMatchingDecoder()); - cli.init(); - int totalCats = 0; - for (int i = 0; i < data.length; i++) { - // doc no. i - Set values = new HashSet(); - for (int j = 0; j < data[i].length; j++) { - values.add(data[i][j]); - } - boolean hasDoc = cli.skipTo(i); - if (hasDoc) { - assertTrue("Document "+i+" must not have a payload!", i==0 || i==2 ); - long cat; - while ((cat = cli.nextCategory()) < Integer.MAX_VALUE) { - assertTrue("expected category not found: " + cat, values.contains((int) cat)); - ++totalCats; - } - } else { - assertFalse("Document "+i+" must have a payload!", i==0 || i==2 ); - } - - } - assertEquals("Wrong number of total categories!", 4, totalCats); - - // Ok.. went through the first 4 docs, now lets try the 6th doc (docid 5) - assertFalse("Doc #6 (docid=5) should not have a payload!",cli.skipTo(5)); - reader.close(); - dir.close(); - } - -}