lucenenet-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From synhers...@apache.org
Subject lucenenet git commit: More porting fixes
Date Sun, 09 Nov 2014 03:32:27 GMT
Repository: lucenenet
Updated Branches:
  refs/heads/master bcff24dd1 -> 6f6f938ef


More porting fixes


Project: http://git-wip-us.apache.org/repos/asf/lucenenet/repo
Commit: http://git-wip-us.apache.org/repos/asf/lucenenet/commit/6f6f938e
Tree: http://git-wip-us.apache.org/repos/asf/lucenenet/tree/6f6f938e
Diff: http://git-wip-us.apache.org/repos/asf/lucenenet/diff/6f6f938e

Branch: refs/heads/master
Commit: 6f6f938ef0ba0cce5079a22782926a9f855f023d
Parents: bcff24d
Author: Itamar Syn-Hershko <itamar@code972.com>
Authored: Sun Nov 9 05:31:49 2014 +0200
Committer: Itamar Syn-Hershko <itamar@code972.com>
Committed: Sun Nov 9 05:32:16 2014 +0200

----------------------------------------------------------------------
 .../CharFilter/MappingCharFilterFactory.cs      | 20 +++---
 .../Analysis/CommonGrams/CommonGramsFilter.cs   | 66 ++++++++------------
 .../CommonGrams/CommonGramsFilterFactory.cs     |  4 +-
 .../CommonGrams/CommonGramsQueryFilter.cs       | 26 +++-----
 .../Analysis/Core/KeywordAnalyzer.cs            |  2 +-
 .../Analysis/Core/WhitespaceAnalyzer.cs         |  2 +-
 .../Analysis/Util/FilteringTokenFilter.cs       |  4 +-
 .../Collation/CollationKeyAnalyzer.cs           | 19 +++---
 src/Lucene.Net.Core/Analysis/Analyzer.cs        |  2 +-
 src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs |  2 +-
 .../Analysis/CachingTokenFilter.cs              |  4 +-
 src/Lucene.Net.Core/Analysis/TokenFilter.cs     | 10 +--
 .../Suggest/Analyzing/SuggestStopFilter.cs      |  6 +-
 .../Analysis/LookaheadTokenFilter.cs            |  2 +-
 .../Analysis/MockAnalyzer.cs                    |  2 +-
 .../Analysis/MockBytesAnalyzer.cs               |  2 +-
 .../Analysis/MockFixedLengthPayloadFilter.cs    |  2 +-
 .../Analysis/MockHoleInjectingTokenFilter.cs    |  2 +-
 .../Analysis/MockPayloadAnalyzer.cs             |  4 +-
 .../Analysis/MockTokenFilter.cs                 |  2 +-
 .../Analysis/MockTokenizer.cs                   |  6 +-
 .../Analysis/MockVariableLengthPayloadFilter.cs |  2 +-
 .../Analysis/ValidatingTokenFilter.cs           |  2 +-
 .../core/Analysis/TestGraphTokenizers.cs        | 18 +++---
 .../core/Analysis/TestLookaheadTokenFilter.cs   |  6 +-
 .../core/Analysis/TestMockAnalyzer.cs           |  4 +-
 .../core/Analysis/TestMockCharFilter.cs         |  2 +-
 .../Codecs/Lucene41/TestBlockPostingsFormat3.cs |  2 +-
 .../Index/TestDocInverterPerFieldErrorInfo.cs   |  2 +-
 .../core/Index/TestDocumentWriter.cs            |  6 +-
 .../core/Index/TestIndexWriter.cs               |  8 +--
 .../core/Index/TestIndexWriterCommit.cs         |  4 +-
 .../core/Index/TestIndexWriterDelete.cs         |  2 +-
 .../core/Index/TestIndexWriterExceptions.cs     | 12 ++--
 .../core/Index/TestLazyProxSkipping.cs          |  2 +-
 .../core/Index/TestMultiLevelSkipList.cs        |  4 +-
 src/Lucene.Net.Tests/core/Index/TestPayloads.cs |  4 +-
 .../core/Index/TestTermVectorsReader.cs         |  2 +-
 .../core/Index/TestTermdocPerf.cs               |  2 +-
 .../core/Search/FuzzyTermOnShortTermsTest.cs    |  2 +-
 .../core/Search/Payloads/PayloadHelper.cs       |  4 +-
 .../Search/Payloads/TestPayloadNearQuery.cs     |  4 +-
 .../Search/Payloads/TestPayloadTermQuery.cs     |  4 +-
 .../core/Search/Spans/TestBasics.cs             |  4 +-
 .../core/Search/Spans/TestPayloadSpans.cs       |  6 +-
 .../core/Search/TestPhraseQuery.cs              |  2 +-
 .../core/Search/TestPositionIncrement.cs        |  2 +-
 .../core/Search/TestTermRangeQuery.cs           |  2 +-
 .../core/Util/TestQueryBuilder.cs               | 10 +--
 49 files changed, 143 insertions(+), 170 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs
index 4489b7c..1d8f008 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs
@@ -1,7 +1,11 @@
 using System.Collections.Generic;
+using System.IO;
 using Lucene.Net.Analysis.Util;
+using Lucene.Net.Support;
+using org.apache.lucene.analysis.charfilter;
+using org.apache.lucene.analysis.util;
 
-namespace org.apache.lucene.analysis.charfilter
+namespace Lucene.Net.Analysis.CharFilter
 {
 
 	/*
@@ -20,15 +24,7 @@ namespace org.apache.lucene.analysis.charfilter
 	 * See the License for the specific language governing permissions and
 	 * limitations under the License.
 	 */
-
-
-	using AbstractAnalysisFactory = AbstractAnalysisFactory;
-	using CharFilterFactory = org.apache.lucene.analysis.util.CharFilterFactory;
-	using MultiTermAwareComponent = org.apache.lucene.analysis.util.MultiTermAwareComponent;
-	using ResourceLoader = org.apache.lucene.analysis.util.ResourceLoader;
-	using ResourceLoaderAware = org.apache.lucene.analysis.util.ResourceLoaderAware;
-
-	/// <summary>
+    /// <summary>
 	/// Factory for <seealso cref="MappingCharFilter"/>. 
 	/// <pre class="prettyprint">
 	/// &lt;fieldType name="text_map" class="solr.TextField" positionIncrementGap="100"&gt;
@@ -80,8 +76,6 @@ namespace org.apache.lucene.analysis.charfilter
 			  wlist.AddRange(lines);
 			}
 		  }
-//JAVA TO C# CONVERTER WARNING: The original Java variable was marked 'final':
-//ORIGINAL LINE: final NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
 		  NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
 		  parseRules(wlist, builder);
 		  normMap = builder.build();
@@ -94,7 +88,7 @@ namespace org.apache.lucene.analysis.charfilter
 		}
 	  }
 
-	  public override Reader create(Reader input)
+	  public override TextReader Create(TextReader input)
 	  {
 		// if the map is null, it means there's actually no mappings... just return the original stream
 		// as there is nothing to do here.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
index 2b97da8..06deccf 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
@@ -1,6 +1,4 @@
-using System.Text;
-
-/*
+/*
  * Licensed to the Apache Software Foundation (ASF) under one or more
  * contributor license agreements.  See the NOTICE file distributed with
  * this work for additional information regarding copyright ownership.
@@ -16,19 +14,14 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
+using System.Text;
+using Lucene.Net.Analysis.Tokenattributes;
+using Lucene.Net.Util;
+using org.apache.lucene.analysis.util;
 
-namespace org.apache.lucene.analysis.commongrams
+namespace Lucene.Net.Analysis.CommonGrams
 {
-
-	using OffsetAttribute = org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
-	using PositionIncrementAttribute = org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute;
-	using CharTermAttribute = org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
-	using PositionLengthAttribute = org.apache.lucene.analysis.tokenattributes.PositionLengthAttribute;
-	using TypeAttribute = org.apache.lucene.analysis.tokenattributes.TypeAttribute;
-	using CharArraySet = org.apache.lucene.analysis.util.CharArraySet;
-	using Version = org.apache.lucene.util.Version;
-
-	/*
+    /*
 	 * TODO: Consider implementing https://issues.apache.org/jira/browse/LUCENE-1688 changes to stop list and associated constructors 
 	 */
 
@@ -98,16 +91,14 @@ namespace org.apache.lucene.analysis.commongrams
 	  /// eliminate the middle bigram "of-the"and save a disk seek and a whole set of
 	  /// position lookups.
 	  /// </summary>
-//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET:
-//ORIGINAL LINE: @Override public boolean incrementToken() throws java.io.IOException
-	  public override bool incrementToken()
+	  public override bool IncrementToken()
 	  {
 		// get the next piece of input
 		if (savedState != null)
 		{
-		  restoreState(savedState);
+		  RestoreState(savedState);
 		  savedState = null;
-		  saveTermBuffer();
+		  SaveTermBuffer();
 		  return true;
 		}
 		else if (!input.incrementToken())
@@ -121,23 +112,21 @@ namespace org.apache.lucene.analysis.commongrams
 		 */
 		if (lastWasCommon || (Common && buffer.Length > 0))
 		{
-		  savedState = captureState();
-		  gramToken();
+		  savedState = CaptureState();
+		  GramToken();
 		  return true;
 		}
 
-		saveTermBuffer();
+		SaveTermBuffer();
 		return true;
 	  }
 
 	  /// <summary>
 	  /// {@inheritDoc}
 	  /// </summary>
-//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET:
-//ORIGINAL LINE: @Override public void reset() throws java.io.IOException
-	  public override void reset()
+	  public override void Reset()
 	  {
-		base.reset();
+		base.Reset();
 		lastWasCommon = false;
 		savedState = null;
 		buffer.Length = 0;
@@ -153,47 +142,46 @@ namespace org.apache.lucene.analysis.commongrams
 	  {
 		  get
 		  {
-			return commonWords != null && commonWords.contains(termAttribute.buffer(), 0, termAttribute.length());
+			return commonWords != null && commonWords.contains(termAttribute.Buffer(), 0, termAttribute.Length);
 		  }
 	  }
 
 	  /// <summary>
 	  /// Saves this information to form the left part of a gram
 	  /// </summary>
-	  private void saveTermBuffer()
+	  private void SaveTermBuffer()
 	  {
 		buffer.Length = 0;
-		buffer.Append(termAttribute.buffer(), 0, termAttribute.length());
+		buffer.Append(termAttribute.Buffer(), 0, termAttribute.Length);
 		buffer.Append(SEPARATOR);
-		lastStartOffset = offsetAttribute.startOffset();
+		lastStartOffset = offsetAttribute.StartOffset();
 		lastWasCommon = Common;
 	  }
 
 	  /// <summary>
 	  /// Constructs a compound token.
 	  /// </summary>
-	  private void gramToken()
+	  private void GramToken()
 	  {
-		buffer.Append(termAttribute.buffer(), 0, termAttribute.length());
-		int endOffset = offsetAttribute.endOffset();
+		buffer.Append(termAttribute.Buffer(), 0, termAttribute.Length);
+		int endOffset = offsetAttribute.EndOffset();
 
-		clearAttributes();
+		ClearAttributes();
 
 		int length = buffer.Length;
-		char[] termText = termAttribute.buffer();
+		char[] termText = termAttribute.Buffer();
 		if (length > termText.Length)
 		{
-		  termText = termAttribute.resizeBuffer(length);
+		  termText = termAttribute.ResizeBuffer(length);
 		}
 
-		buffer.getChars(0, length, termText, 0);
+		buffer.GetChars(0, length, termText, 0);
 		termAttribute.Length = length;
 		posIncAttribute.PositionIncrement = 0;
 		posLenAttribute.PositionLength = 2; // bigram
-		offsetAttribute.setOffset(lastStartOffset, endOffset);
+		offsetAttribute.SetOffset(lastStartOffset, endOffset);
 		typeAttribute.Type = GRAM_TYPE;
 		buffer.Length = 0;
 	  }
 	}
-
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs
index 2233e83..fe531cb 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs
@@ -92,9 +92,9 @@ namespace Lucene.Net.Analysis.CommonGrams
 		  }
 	  }
 
-	  public override TokenFilter Create(TokenStream input)
+	  public override TokenStream Create(TokenStream input)
 	  {
-		CommonGramsFilter commonGrams = new CommonGramsFilter(luceneMatchVersion, input, commonWords);
+		var commonGrams = new CommonGramsFilter(luceneMatchVersion, input, commonWords);
 		return commonGrams;
 	  }
 	}

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs
index b787bde..e708697 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs
@@ -14,15 +14,11 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-namespace org.apache.lucene.analysis.commongrams
-{
-
-	using PositionIncrementAttribute = org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute;
-	using TypeAttribute = org.apache.lucene.analysis.tokenattributes.TypeAttribute;
 
-//JAVA TO C# CONVERTER TODO TASK: This Java 'import static' statement cannot be converted to C#:
-//	import static org.apache.lucene.analysis.commongrams.CommonGramsFilter.GRAM_TYPE;
+using Lucene.Net.Analysis.Tokenattributes;
 
+namespace Lucene.Net.Analysis.CommonGrams
+{
 	/// <summary>
 	/// Wrap a CommonGramsFilter optimizing phrase queries by only returning single
 	/// words when they are not a member of a bigram.
@@ -62,11 +58,9 @@ namespace org.apache.lucene.analysis.commongrams
 	  /// <summary>
 	  /// {@inheritDoc}
 	  /// </summary>
-//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET:
-//ORIGINAL LINE: @Override public void reset() throws java.io.IOException
-	  public override void reset()
+	  public override void Reset()
 	  {
-		base.reset();
+		base.Reset();
 		previous = null;
 		previousType = null;
 		exhausted = false;
@@ -80,17 +74,15 @@ namespace org.apache.lucene.analysis.commongrams
 	  /// <li>output:"the-rain", "rain-in" ,"in-spain", "falls", "mainly"
 	  /// </ul>
 	  /// </summary>
-//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET:
-//ORIGINAL LINE: @Override public boolean incrementToken() throws java.io.IOException
-	  public override bool incrementToken()
+	  public override bool IncrementToken()
 	  {
 		while (!exhausted && input.incrementToken())
 		{
-		  State current = captureState();
+		  State current = CaptureState();
 
 		  if (previous != null && !GramType)
 		  {
-			restoreState(previous);
+			RestoreState(previous);
 			previous = current;
 			previousType = typeAttribute.type();
 
@@ -111,7 +103,7 @@ namespace org.apache.lucene.analysis.commongrams
 		  return false;
 		}
 
-		restoreState(previous);
+		RestoreState(previous);
 		previous = null;
 
 		if (GramType)

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs
index 1f1a42b..4d13021 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs
@@ -32,7 +32,7 @@ namespace Lucene.Net.Analysis.Core
         {
         }
 
-        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+        public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             return new TokenStreamComponents(new KeywordTokenizer(reader));
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs
index 7e77c8d..4ef12c7 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs
@@ -50,7 +50,7 @@ namespace Lucene.Net.Analysis.Core
             this.matchVersion = matchVersion;
         }
 
-        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+        public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             return new TokenStreamComponents(new WhitespaceTokenizer(matchVersion, reader));
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs
index 4d55a25..0feb6b8 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs
@@ -81,7 +81,7 @@ namespace Lucene.Net.Analysis.Util
             if (enablePositionIncrements)
             {
                 skippedPositions = 0;
-                while (Input.IncrementToken())
+                while (input.IncrementToken())
                 {
                     if (Accept())
                     {
@@ -96,7 +96,7 @@ namespace Lucene.Net.Analysis.Util
             }
             else
             {
-                while (Input.IncrementToken())
+                while (input.IncrementToken())
                 {
                     if (Accept())
                     {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
index 06fb9e0..0c4df1f 100644
--- a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
+++ b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs
@@ -1,8 +1,10 @@
 using System;
+using System.IO;
+using Lucene.Net.Analysis;
 using Lucene.Net.Analysis.Core;
-using Lucene.Net.Collation;
+using org.apache.lucene.collation;
 
-namespace org.apache.lucene.collation
+namespace Lucene.Net.Collation
 {
 
 	/*
@@ -23,13 +25,10 @@ namespace org.apache.lucene.collation
 	 */
 
 
-	using Analyzer = org.apache.lucene.analysis.Analyzer;
-	using KeywordTokenizer = KeywordTokenizer;
-	using IndexableBinaryStringTools = org.apache.lucene.util.IndexableBinaryStringTools; // javadoc @link
-	using Version = org.apache.lucene.util.Version;
+    // javadoc @link
 
 
-	/// <summary>
+    /// <summary>
 	/// <para>
 	///   Configures <seealso cref="KeywordTokenizer"/> with <seealso cref="CollationAttributeFactory"/>.
 	/// </para>
@@ -106,14 +105,14 @@ namespace org.apache.lucene.collation
 
 	  /// @deprecated Use <seealso cref="CollationKeyAnalyzer#CollationKeyAnalyzer(Version, Collator)"/>
 	  ///   and specify a version instead. This ctor will be removed in Lucene 5.0 
-	  [Obsolete("Use <seealso cref="CollationKeyAnalyzer#CollationKeyAnalyzer(org.apache.lucene.util.Version, java.text.Collator)"/>")]
+	  [Obsolete("Use <seealso cref=\"CollationKeyAnalyzer#CollationKeyAnalyzer(org.apache.lucene.util.Version, java.text.Collator)\"/>")]
 	  public CollationKeyAnalyzer(Collator collator) : this(Version.LUCENE_31, collator)
 	  {
 	  }
 
-	  protected internal override TokenStreamComponents createComponents(string fieldName, Reader reader)
+        public override Analyzer.TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
 	  {
-		if (matchVersion.onOrAfter(Version.LUCENE_40))
+		if (matchVersion.OnOrAfter(Version.LUCENE_40))
 		{
 		  KeywordTokenizer tokenizer = new KeywordTokenizer(factory, reader, KeywordTokenizer.DEFAULT_BUFFER_SIZE);
 		  return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/Analyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/Analyzer.cs b/src/Lucene.Net.Core/Analysis/Analyzer.cs
index 0ae0c8e..c183885 100644
--- a/src/Lucene.Net.Core/Analysis/Analyzer.cs
+++ b/src/Lucene.Net.Core/Analysis/Analyzer.cs
@@ -104,7 +104,7 @@ namespace Lucene.Net.Analysis
         /// <param name="reader">
         ///          the reader passed to the <seealso cref="Tokenizer"/> constructor </param>
         /// <returns> the <seealso cref="TokenStreamComponents"/> for this analyzer. </returns>
-        protected abstract TokenStreamComponents CreateComponents(string fieldName, TextReader reader);
+        public abstract TokenStreamComponents CreateComponents(string fieldName, TextReader reader);
 
         /// <summary>
         /// Returns a TokenStream suitable for <code>fieldName</code>, tokenizing

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs b/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs
index f58467b..a5e8660 100644
--- a/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs
+++ b/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs
@@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis
             return reader;
         }
 
-        protected override sealed TokenStreamComponents CreateComponents(string fieldName, TextReader aReader)
+        public override sealed TokenStreamComponents CreateComponents(string fieldName, TextReader aReader)
         {
             return WrapComponents(fieldName, GetWrappedAnalyzer(fieldName).CreateComponents(fieldName, aReader));
         }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs b/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs
index 77994aa..3a1706d 100644
--- a/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs
+++ b/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs
@@ -91,12 +91,12 @@ namespace Lucene.Net.Analysis
 
         private void FillCache()
         {
-            while (Input.IncrementToken())
+            while (input.IncrementToken())
             {
                 Cache.AddLast(CaptureState());
             }
             // capture final state
-            Input.End();
+            input.End();
             FinalState = CaptureState();
         }
     }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/TokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Core/Analysis/TokenFilter.cs b/src/Lucene.Net.Core/Analysis/TokenFilter.cs
index 70f1373..6ed98a5 100644
--- a/src/Lucene.Net.Core/Analysis/TokenFilter.cs
+++ b/src/Lucene.Net.Core/Analysis/TokenFilter.cs
@@ -26,14 +26,14 @@ namespace Lucene.Net.Analysis
     {
         /// <summary>
         /// The source of tokens for this filter. </summary>
-        protected internal readonly TokenStream Input;
+        protected internal readonly TokenStream input;
 
         /// <summary>
         /// Construct a token stream filtering the given input. </summary>
         protected internal TokenFilter(TokenStream input)
             : base(input)
         {
-            this.Input = input;
+            this.input = input;
         }
 
         /// <summary>
@@ -45,7 +45,7 @@ namespace Lucene.Net.Analysis
         /// </summary>
         public override void End()
         {
-            Input.End();
+            input.End();
         }
 
         /// <summary>
@@ -57,7 +57,7 @@ namespace Lucene.Net.Analysis
         /// </summary>
         public override void Dispose()
         {
-            Input.Dispose();
+            input.Dispose();
         }
 
         /// <summary>
@@ -69,7 +69,7 @@ namespace Lucene.Net.Analysis
         /// </summary>
         public override void Reset()
         {
-            Input.Reset();
+            input.Reset();
         }
     }
 }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs b/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs
index 3f2f72c..1451ac9 100644
--- a/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs
+++ b/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs
@@ -83,7 +83,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
                 return false;
             }
 
-            if (!Input.IncrementToken())
+            if (!input.IncrementToken())
             {
                 return false;
             }
@@ -97,7 +97,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
                     int endOffset = offsetAtt.EndOffset();
                     // This token may be a stopword, if it's not end:
                     State sav = CaptureState();
-                    if (Input.IncrementToken())
+                    if (input.IncrementToken())
                     {
                         // It was a stopword; skip it
                         skippedPositions += posInc;
@@ -105,7 +105,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing
                     else
                     {
                         ClearAttributes();
-                        Input.End();
+                        input.End();
                         endState = CaptureState();
                         int finalEndOffset = offsetAtt.EndOffset();
                         Debug.Assert(finalEndOffset >= endOffset);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs
index 0bc4bbe..9b2aee1 100644
--- a/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs
@@ -180,7 +180,7 @@ namespace Lucene.Net.Analysis
                 positions.Get(InputPos).Add(CaptureState());
                 TokenPending = false;
             }
-            bool gotToken = Input.IncrementToken();
+            bool gotToken = input.IncrementToken();
             if (DEBUG)
             {
                 Console.WriteLine("  input.incrToken() returned " + gotToken);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs
index 37fe8cd..9a1b883 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs
@@ -89,7 +89,7 @@ namespace Lucene.Net.Analysis
         {
         }
 
-        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+        public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             MockTokenizer tokenizer = new MockTokenizer(reader, RunAutomaton, LowerCase, MaxTokenLength_Renamed);
             tokenizer.EnableChecks = EnableChecks_Renamed;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs
index 12934df..6967b75 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs
@@ -26,7 +26,7 @@ namespace Lucene.Net.Analysis
     {
         private readonly MockBytesAttributeFactory Factory = new MockBytesAttributeFactory();
 
-        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+        public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             Tokenizer t = new MockTokenizer(Factory, reader, MockTokenizer.KEYWORD, false, MockTokenizer.DEFAULT_MAX_TOKEN_LENGTH);
             return new TokenStreamComponents(t);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs
index 7b510ba..d4a6646 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs
@@ -47,7 +47,7 @@ namespace Lucene.Net.Analysis
 
         public override bool IncrementToken()
         {
-            if (Input.IncrementToken())
+            if (input.IncrementToken())
             {
                 Random.NextBytes((byte[])(Array)Bytes);
                 PayloadAtt.Payload = Payload;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs
index c5c2577..fceb767 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs
@@ -56,7 +56,7 @@ namespace Lucene.Net.Analysis
 
         public override bool IncrementToken()
         {
-            if (Input.IncrementToken())
+            if (input.IncrementToken())
             {
                 int posInc = PosIncAtt.PositionIncrement;
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs
index 60ee80e..b348670 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs
@@ -31,7 +31,7 @@ namespace Lucene.Net.Analysis
     /// </summary>
     public sealed class MockPayloadAnalyzer : Analyzer
     {
-        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+        public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
         {
             Tokenizer result = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
             return new TokenStreamComponents(result, new MockPayloadFilter(result, fieldName));
@@ -66,7 +66,7 @@ namespace Lucene.Net.Analysis
 
         public override bool IncrementToken()
         {
-            if (Input.IncrementToken())
+            if (input.IncrementToken())
             {
                 PayloadAttr.Payload = new BytesRef(("pos: " + Pos)/*.getBytes(IOUtils.CHARSET_UTF_8)*/);
                 int posIncr;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs
index f8467dc..f692726 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs
@@ -69,7 +69,7 @@ namespace Lucene.Net.Analysis
 
             // return the first non-stop word found
             SkippedPositions = 0;
-            while (Input.IncrementToken())
+            while (input.IncrementToken())
             {
                 if (!Filter.Run(TermAtt.Buffer(), 0, TermAtt.Length))
                 {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs b/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs
index 5fa4ae3..7cb9f28 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs
@@ -247,14 +247,14 @@ namespace Lucene.Net.Analysis
                     {
                         // read(char[])
                         char[] c = new char[1];
-                        int ret = Input.Read(c, 0, c.Length);
+                        int ret = input.Read(c, 0, c.Length);
                         return ret == 0 ? -1 : c[0];
                     }
                 case 1:
                     {
                         // read(char[], int, int)
                         char[] c = new char[2];
-                        int ret = Input.Read(c, 1, 1);
+                        int ret = input.Read(c, 1, 1);
                         return ret == 0 ? -1 : c[1];
                     }
                 /* LUCENE TO-DO not sure if needed, CharBuffer not supported
@@ -268,7 +268,7 @@ namespace Lucene.Net.Analysis
                   }*/
                 default:
                     // read()
-                    return Input.Read();
+                    return input.Read();
             }
         }
 

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs
index 40f2151..07631bf 100644
--- a/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs
@@ -44,7 +44,7 @@ namespace Lucene.Net.Analysis
 
         public override bool IncrementToken()
         {
-            if (Input.IncrementToken())
+            if (input.IncrementToken())
             {
                 Random.NextBytes((byte[])(Array)Bytes);
                 Payload.Length = Random.Next(MAXLENGTH);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs
index 3970a3e..bb5cee6 100644
--- a/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs
+++ b/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs
@@ -99,7 +99,7 @@ namespace Lucene.Net.Analysis
 
         public override bool IncrementToken()
         {
-            if (!Input.IncrementToken())
+            if (!input.IncrementToken())
             {
                 return false;
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs b/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs
index 4fe19b1..bf85d9a 100644
--- a/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs
+++ b/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs
@@ -204,7 +204,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t2 = new MockGraphTokenFilter(Random(), t);
@@ -239,7 +239,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new GraphTokenizer(reader);
                 TokenStream t2 = new MockGraphTokenFilter(Random(), t);
@@ -278,7 +278,7 @@ namespace Lucene.Net.Analysis
             {
                 while (true)
                 {
-                    bool gotOne = Input.IncrementToken();
+                    bool gotOne = input.IncrementToken();
                     if (!gotOne)
                     {
                         return false;
@@ -328,7 +328,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t2 = new MockGraphTokenFilter(Random(), t);
@@ -368,7 +368,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t2 = new RemoveATokens(t);
@@ -405,7 +405,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t2 = new MockGraphTokenFilter(Random(), t);
@@ -449,7 +449,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t1 = new MockGraphTokenFilter(Random(), t);
@@ -486,7 +486,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t1 = new MockGraphTokenFilter(Random(), t);
@@ -523,7 +523,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TokenStream t1 = new MockHoleInjectingTokenFilter(Random(), t);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs b/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs
index db48b98..04622b8 100644
--- a/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs
+++ b/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs
@@ -41,7 +41,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Random random = Random();
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, random.NextBoolean());
@@ -84,7 +84,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, Random().NextBoolean());
                 TokenStream output = new NeverPeeksLookaheadTokenFilter(tokenizer);
@@ -109,7 +109,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 TrivialLookaheadFilter filter = new TrivialLookaheadFilter(source);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs b/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs
index e39be6a..e185484 100644
--- a/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs
+++ b/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs
@@ -182,7 +182,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false, 5);
                 return new TokenStreamComponents(t, t);
@@ -259,7 +259,7 @@ namespace Lucene.Net.Analysis
                 this.Limit = limit;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader, Dfa, Lowercase, Limit);
                 return new TokenStreamComponents(t, t);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs b/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs
index cabd3f2..03df580 100644
--- a/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs
+++ b/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs
@@ -44,7 +44,7 @@ namespace Lucene.Net.Analysis
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs b/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs
index ca67a17..138d941 100644
--- a/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs
+++ b/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs
@@ -150,7 +150,7 @@ namespace Lucene.Net.Codecs.Lucene41
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader);
                 if (fieldName.Contains("payloadsFixed"))

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs b/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs
index e2eb471..8ec9e62 100644
--- a/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs
@@ -54,7 +54,7 @@ namespace Lucene.Net.Index
 
         private class ThrowingAnalyzer : Analyzer
         {
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader input)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader input)
             {
                 Tokenizer tokenizer = new MockTokenizer(input);
                 if (fieldName.Equals("distinctiveFieldName"))

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs b/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs
index e66a2fe..997f4ff 100644
--- a/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs
@@ -146,7 +146,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, false));
             }
@@ -195,7 +195,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new TokenFilterAnonymousInnerClassHelper(this, tokenizer));
@@ -230,7 +230,7 @@ namespace Lucene.Net.Index
                         return true;
                     }
 
-                    bool hasNext = Input.IncrementToken();
+                    bool hasNext = input.IncrementToken();
                     if (!hasNext)
                     {
                         return false;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs
index a9801f3..8631da7 100644
--- a/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs
@@ -1866,7 +1866,7 @@ namespace Lucene.Net.Index
 
         internal sealed class StringSplitAnalyzer : Analyzer
         {
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new StringSplitTokenizer(reader));
             }
@@ -2226,7 +2226,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 throw new InvalidOperationException("don't invoke me!");
             }
@@ -2301,7 +2301,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader);
                 TokenStream stream = new MockTokenFilter(tokenizer, MockTokenFilter.ENGLISH_STOPSET);
@@ -2346,7 +2346,7 @@ namespace Lucene.Net.Index
                 this.SecondSet = secondSet;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader);
                 TokenStream stream = new MockTokenFilter(tokenizer, MockTokenFilter.ENGLISH_STOPSET);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs
index 1a4b4d3..4580112 100644
--- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs
@@ -258,7 +258,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true));
             }
@@ -276,7 +276,7 @@ namespace Lucene.Net.Index
                 this.Length = length;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
                 return new TokenStreamComponents(tokenizer, new MockFixedLengthPayloadFilter(Random(), tokenizer, Length));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs
index d5f8513..60dff7e 100644
--- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs
@@ -1192,7 +1192,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs
index 58b706e..6fc21f6 100644
--- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs
@@ -448,7 +448,7 @@ namespace Lucene.Net.Index
                 {
                     throw new IOException(CRASH_FAIL_MESSAGE);
                 }
-                return Input.IncrementToken();
+                return input.IncrementToken();
             }
 
             public override void Reset()
@@ -519,7 +519,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases.
@@ -637,7 +637,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases.
@@ -663,7 +663,7 @@ namespace Lucene.Net.Index
                     {
                         throw new IOException();
                     }
-                    return Input.IncrementToken();
+                    return input.IncrementToken();
                 }
 
                 public override void Reset()
@@ -873,7 +873,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases.
@@ -972,7 +972,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false);
                 tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs b/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs
index 59e3e02..4010ec4 100644
--- a/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs
@@ -121,7 +121,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs b/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs
index 3caf34c..aa94dc8 100644
--- a/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs
@@ -127,7 +127,7 @@ namespace Lucene.Net.Index
         {
             internal readonly AtomicInteger PayloadCount = new AtomicInteger(-1);
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true);
                 return new TokenStreamComponents(tokenizer, new PayloadFilter(PayloadCount, tokenizer));
@@ -148,7 +148,7 @@ namespace Lucene.Net.Index
 
             public override bool IncrementToken()
             {
-                bool hasNext = Input.IncrementToken();
+                bool hasNext = input.IncrementToken();
                 if (hasNext)
                 {
                     PayloadAtt.Payload = new BytesRef(new sbyte[] { (sbyte)PayloadCount.IncrementAndGet() });

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestPayloads.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestPayloads.cs b/src/Lucene.Net.Tests/core/Index/TestPayloads.cs
index 903225a..dde563c 100644
--- a/src/Lucene.Net.Tests/core/Index/TestPayloads.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestPayloads.cs
@@ -387,7 +387,7 @@ namespace Lucene.Net.Index
                 FieldToData[field] = new PayloadData(data, offset, length);
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 PayloadData payload;
                 FieldToData.TryGetValue(fieldName, out payload);
@@ -436,7 +436,7 @@ namespace Lucene.Net.Index
 
             public override bool IncrementToken()
             {
-                bool hasNext = Input.IncrementToken();
+                bool hasNext = input.IncrementToken();
                 if (!hasNext)
                 {
                     return false;

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs b/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs
index 24deec6..7995318 100644
--- a/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs
@@ -228,7 +228,7 @@ namespace Lucene.Net.Index
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MyTokenizer(OuterInstance, reader));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs b/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs
index fe83c63..5160a75 100644
--- a/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs
+++ b/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs
@@ -119,7 +119,7 @@ namespace Lucene.Net.Index
                 this.PercentDocs = percentDocs;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new RepeatingTokenizer(reader, Val, Random, PercentDocs, MaxTF));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs b/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs
index 1b9fb84..b668aad 100644
--- a/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs
+++ b/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs
@@ -89,7 +89,7 @@ namespace Lucene.Net.Search
             {
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs b/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs
index a88b90e..71ea1e1 100644
--- a/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs
+++ b/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs
@@ -63,7 +63,7 @@ namespace Lucene.Net.Search.Payloads
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result, fieldName));
@@ -88,7 +88,7 @@ namespace Lucene.Net.Search.Payloads
 
             public override bool IncrementToken()
             {
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     if (FieldName.Equals(FIELD))
                     {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs
index 9df0229..82ef6d1 100644
--- a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs
+++ b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs
@@ -51,7 +51,7 @@ namespace Lucene.Net.Search.Payloads
 
         private class PayloadAnalyzer : Analyzer
         {
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(result, new PayloadFilter(result, fieldName));
@@ -74,7 +74,7 @@ namespace Lucene.Net.Search.Payloads
             public override bool IncrementToken()
             {
                 bool result = false;
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     if (NumSeen % 2 == 0)
                     {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs
index 2f0ced7..8f49550 100644
--- a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs
+++ b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs
@@ -58,7 +58,7 @@ namespace Lucene.Net.Search.Payloads
             {
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(result, new PayloadFilter(result, fieldName));
@@ -81,7 +81,7 @@ namespace Lucene.Net.Search.Payloads
 
             public override bool IncrementToken()
             {
-                bool hasNext = Input.IncrementToken();
+                bool hasNext = input.IncrementToken();
                 if (hasNext)
                 {
                     if (FieldName.Equals("field"))

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs b/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs
index 0aef0b3..b013b58 100644
--- a/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs
+++ b/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs
@@ -72,7 +72,7 @@ namespace Lucene.Net.Search.Spans
 
             public override bool IncrementToken()
             {
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     PayloadAttr.Payload = new BytesRef(("pos: " + Pos).GetBytes(IOUtils.CHARSET_UTF_8));
                     Pos++;
@@ -118,7 +118,7 @@ namespace Lucene.Net.Search.Spans
             {
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, new SimplePayloadFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs b/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs
index 45ce22a..6017223 100644
--- a/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs
+++ b/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs
@@ -510,7 +510,7 @@ namespace Lucene.Net.Search.Spans
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result));
@@ -544,7 +544,7 @@ namespace Lucene.Net.Search.Spans
 
             public override bool IncrementToken()
             {
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     string token = TermAtt.ToString();
 
@@ -581,7 +581,7 @@ namespace Lucene.Net.Search.Spans
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs b/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs
index 2e1ac1d..de56994 100644
--- a/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs
+++ b/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs
@@ -90,7 +90,7 @@ namespace Lucene.Net.Search
             {
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, false));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs b/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs
index 83b0970..982c244 100644
--- a/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs
+++ b/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs
@@ -173,7 +173,7 @@ namespace Lucene.Net.Search
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new TokenizerAnonymousInnerClassHelper(this, reader));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs b/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs
index e6541b1..54054e0 100644
--- a/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs
+++ b/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs
@@ -259,7 +259,7 @@ namespace Lucene.Net.Search
                 }
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new SingleCharTokenizer(reader));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs b/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs
index 0140fc0..421a203 100644
--- a/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs
+++ b/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs
@@ -135,7 +135,7 @@ namespace Lucene.Net.Util
         /// adds synonym of "dog" for "dogs". </summary>
         internal class MockSynonymAnalyzer : Analyzer
         {
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new MockSynonymFilter(tokenizer));
@@ -169,7 +169,7 @@ namespace Lucene.Net.Util
                     return true;
                 }
 
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     AddSynonym = TermAtt.ToString().Equals("dogs");
                     return true;
@@ -240,7 +240,7 @@ namespace Lucene.Net.Util
                 this.OuterInstance = outerInstance;
             }
 
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new SimpleCJKTokenizer(reader));
             }
@@ -316,7 +316,7 @@ namespace Lucene.Net.Util
                     return true;
                 }
 
-                if (Input.IncrementToken())
+                if (input.IncrementToken())
                 {
                     AddSynonym = TermAtt.ToString().Equals("国");
                     return true;
@@ -330,7 +330,7 @@ namespace Lucene.Net.Util
 
         internal class MockCJKSynonymAnalyzer : Analyzer
         {
-            protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
+            public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new SimpleCJKTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new MockCJKSynonymFilter(tokenizer));


Mime
View raw message