--- /dev/null
+package org.apache.lucene.index;
+
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+import java.io.IOException;
+
+import org.apache.lucene.analysis.MockAnalyzer;
+import org.apache.lucene.analysis.TokenStream;
+import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
+import org.apache.lucene.document.Document;
+import org.apache.lucene.document.Field;
+import org.apache.lucene.index.FieldInfo.IndexOptions;
+import org.apache.lucene.store.MockDirectoryWrapper;
+import org.apache.lucene.util.LuceneTestCase;
+import org.apache.lucene.util._TestUtil;
+import org.junit.Ignore;
+
+/**
+ * Test indexes ~82M docs with 26 terms each, so you get > Integer.MAX_VALUE terms/docs pairs
+ * @lucene.experimental
+ */
+public class Test2BPostings extends LuceneTestCase {
+
+ @Nightly
+ public void test() throws Exception {
+
+ MockDirectoryWrapper dir = newFSDirectory(_TestUtil.getTempDir("2BPostings"));
+ dir.setThrottling(MockDirectoryWrapper.Throttling.NEVER);
+ dir.setCheckIndexOnClose(false); // don't double-checkindex
+
+ IndexWriter w = new IndexWriter(dir,
+ new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))
+ .setMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
+ .setRAMBufferSizeMB(256.0)
+ .setMergeScheduler(new ConcurrentMergeScheduler())
+ .setMergePolicy(newLogMergePolicy(false, 10))
+ .setOpenMode(IndexWriterConfig.OpenMode.CREATE));
+
+ MergePolicy mp = w.getConfig().getMergePolicy();
+ if (mp instanceof LogByteSizeMergePolicy) {
+ // 1 petabyte:
+ ((LogByteSizeMergePolicy) mp).setMaxMergeMB(1024*1024*1024);
+ }
+
+ Document doc = new Document();
+ Field field = new Field("field", new MyTokenStream());
+ field.setIndexOptions(IndexOptions.DOCS_ONLY);
+ field.setOmitNorms(true);
+ doc.add(field);
+
+ final int numDocs = (Integer.MAX_VALUE / 26) + 1;
+ for (int i = 0; i < numDocs; i++) {
+ w.addDocument(doc);
+ if (VERBOSE && i % 100000 == 0) {
+ System.out.println(i + " of " + numDocs + "...");
+ }
+ }
+ w.optimize();
+ w.close();
+ CheckIndex ci = new CheckIndex(dir);
+ if (VERBOSE) {
+ ci.setInfoStream(System.out);
+ }
+ ci.checkIndex();
+ dir.close();
+ }
+
+ public static final class MyTokenStream extends TokenStream {
+ private final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
+ private final char buffer[];
+ int index;
+
+ public MyTokenStream() {
+ termAtt.setLength(1);
+ buffer = termAtt.buffer();
+ }
+
+ @Override
+ public boolean incrementToken() throws IOException {
+ if (index <= 'z') {
+ buffer[0] = (char) index++;
+ return true;
+ }
+ return false;
+ }
+
+ @Override
+ public void reset() throws IOException {
+ index = 'a';
+ }
+ }
+}