lucenenet-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ccurr...@apache.org
Subject svn commit: r1303294 [1/6] - in /incubator/lucene.net/trunk: src/contrib/Analyzers/AR/ src/contrib/Analyzers/BR/ src/contrib/Analyzers/CJK/ src/contrib/Analyzers/Cn/ src/contrib/Analyzers/Compound/ src/contrib/Analyzers/Cz/ src/contrib/Analyzers/De/ sr...
Date Wed, 21 Mar 2012 06:04:35 GMT
Author: ccurrens
Date: Wed Mar 21 06:04:26 2012
New Revision: 1303294

URL: http://svn.apache.org/viewvc?rev=1303294&view=rev
Log:
[LUCENENET-467] - additional work on properties.  Added some and reverted some.  Still need to look through others to make sure they're acting as data.

Modified:
    incubator/lucene.net/trunk/src/contrib/Analyzers/AR/ArabicAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/BR/BrazilianAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKTokenizer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Cn/ChineseAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Compound/CompoundWordTokenFilterBase.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Cz/CzechAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/De/GermanAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/El/GreekAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Fa/PersianAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Fr/FrenchAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAndSuffixAwareTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAwareTokenStream.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/EdgeNGramTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/NGramTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Nl/DutchAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/NumericPayloadTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Query/QueryAutoStopWordAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Ru/RussianAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleAnalyzerWrapper.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleMatrixFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Sinks/TokenTypeSinkFilter.cs
    incubator/lucene.net/trunk/src/contrib/Analyzers/Th/ThaiAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/Core/Index/TermVectorEnumerator.cs
    incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/FieldTermStack.cs
    incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/VectorHighlightMapper.cs
    incubator/lucene.net/trunk/src/contrib/Highlighter/QueryTermExtractor.cs
    incubator/lucene.net/trunk/src/contrib/Highlighter/SimpleFragmenter.cs
    incubator/lucene.net/trunk/src/contrib/Highlighter/TokenGroup.cs
    incubator/lucene.net/trunk/src/contrib/Highlighter/TokenSources.cs
    incubator/lucene.net/trunk/src/contrib/Queries/FuzzyLikeThisQuery.cs
    incubator/lucene.net/trunk/src/contrib/Queries/Similar/MoreLikeThis.cs
    incubator/lucene.net/trunk/src/contrib/Similarity/Similar/MoreLikeThis.cs
    incubator/lucene.net/trunk/src/contrib/Snowball/Lucene.Net/Analysis/Snowball/SnowballAnalyzer.cs
    incubator/lucene.net/trunk/src/contrib/SpellChecker/Spell/SpellChecker.cs
    incubator/lucene.net/trunk/src/core/Analysis/Analyzer.cs
    incubator/lucene.net/trunk/src/core/Analysis/KeywordAnalyzer.cs
    incubator/lucene.net/trunk/src/core/Analysis/NumericTokenStream.cs
    incubator/lucene.net/trunk/src/core/Analysis/PerFieldAnalyzerWrapper.cs
    incubator/lucene.net/trunk/src/core/Analysis/PorterStemFilter.cs
    incubator/lucene.net/trunk/src/core/Analysis/PorterStemmer.cs
    incubator/lucene.net/trunk/src/core/Analysis/SimpleAnalyzer.cs
    incubator/lucene.net/trunk/src/core/Analysis/Standard/StandardAnalyzer.cs
    incubator/lucene.net/trunk/src/core/Analysis/Standard/StandardFilter.cs
    incubator/lucene.net/trunk/src/core/Analysis/Standard/StandardTokenizer.cs
    incubator/lucene.net/trunk/src/core/Analysis/StopAnalyzer.cs
    incubator/lucene.net/trunk/src/core/Analysis/TeeSinkTokenFilter.cs
    incubator/lucene.net/trunk/src/core/Analysis/Token.cs
    incubator/lucene.net/trunk/src/core/Analysis/Tokenattributes/OffsetAttribute.cs
    incubator/lucene.net/trunk/src/core/Analysis/Tokenattributes/OffsetAttributeImpl.cs
    incubator/lucene.net/trunk/src/core/Analysis/Tokenattributes/TypeAttribute.cs
    incubator/lucene.net/trunk/src/core/Analysis/Tokenattributes/TypeAttributeImpl.cs
    incubator/lucene.net/trunk/src/core/Analysis/WhitespaceAnalyzer.cs
    incubator/lucene.net/trunk/src/core/Document/AbstractField.cs
    incubator/lucene.net/trunk/src/core/Document/Document.cs
    incubator/lucene.net/trunk/src/core/Document/Fieldable.cs
    incubator/lucene.net/trunk/src/core/Index/AbstractAllTermDocs.cs
    incubator/lucene.net/trunk/src/core/Index/CheckIndex.cs
    incubator/lucene.net/trunk/src/core/Index/CompoundFileReader.cs
    incubator/lucene.net/trunk/src/core/Index/CompoundFileWriter.cs
    incubator/lucene.net/trunk/src/core/Index/ConcurrentMergeScheduler.cs
    incubator/lucene.net/trunk/src/core/Index/DirectoryReader.cs
    incubator/lucene.net/trunk/src/core/Index/DocInverterPerField.cs
    incubator/lucene.net/trunk/src/core/Index/DocumentsWriter.cs
    incubator/lucene.net/trunk/src/core/Index/FieldsWriter.cs
    incubator/lucene.net/trunk/src/core/Index/FilterIndexReader.cs
    incubator/lucene.net/trunk/src/core/Index/IndexReader.cs
    incubator/lucene.net/trunk/src/core/Index/IndexWriter.cs
    incubator/lucene.net/trunk/src/core/Index/LogMergePolicy.cs
    incubator/lucene.net/trunk/src/core/Index/MultiReader.cs
    incubator/lucene.net/trunk/src/core/Index/ParallelReader.cs
    incubator/lucene.net/trunk/src/core/Index/PositionBasedTermVectorMapper.cs
    incubator/lucene.net/trunk/src/core/Index/SegmentInfo.cs
    incubator/lucene.net/trunk/src/core/Index/SegmentInfos.cs
    incubator/lucene.net/trunk/src/core/Index/SegmentMerger.cs
    incubator/lucene.net/trunk/src/core/Index/SegmentReader.cs
    incubator/lucene.net/trunk/src/core/Index/SegmentTermVector.cs
    incubator/lucene.net/trunk/src/core/Index/SortedTermVectorMapper.cs
    incubator/lucene.net/trunk/src/core/Index/TermFreqVector.cs
    incubator/lucene.net/trunk/src/core/Index/TermPositionVector.cs
    incubator/lucene.net/trunk/src/core/Index/TermVectorEntry.cs
    incubator/lucene.net/trunk/src/core/Index/TermVectorMapper.cs
    incubator/lucene.net/trunk/src/core/Index/TermVectorsReader.cs
    incubator/lucene.net/trunk/src/core/Index/TermVectorsTermsWriterPerField.cs
    incubator/lucene.net/trunk/src/core/Index/TermVectorsWriter.cs
    incubator/lucene.net/trunk/src/core/Messages/Message.cs
    incubator/lucene.net/trunk/src/core/Messages/MessageImpl.cs
    incubator/lucene.net/trunk/src/core/QueryParser/CharStream.cs
    incubator/lucene.net/trunk/src/core/QueryParser/FastCharStream.cs
    incubator/lucene.net/trunk/src/core/QueryParser/QueryParser.cs
    incubator/lucene.net/trunk/src/core/QueryParser/QueryParserTokenManager.cs
    incubator/lucene.net/trunk/src/core/Search/BooleanClause.cs
    incubator/lucene.net/trunk/src/core/Search/BooleanQuery.cs
    incubator/lucene.net/trunk/src/core/Search/BooleanScorer.cs
    incubator/lucene.net/trunk/src/core/Search/Collector.cs
    incubator/lucene.net/trunk/src/core/Search/ConstantScoreQuery.cs
    incubator/lucene.net/trunk/src/core/Search/DisjunctionMaxQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Explanation.cs
    incubator/lucene.net/trunk/src/core/Search/FieldComparator.cs
    incubator/lucene.net/trunk/src/core/Search/FilterManager.cs
    incubator/lucene.net/trunk/src/core/Search/FilteredQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Function/CustomScoreQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Function/ValueSourceQuery.cs
    incubator/lucene.net/trunk/src/core/Search/FuzzyQuery.cs
    incubator/lucene.net/trunk/src/core/Search/IndexSearcher.cs
    incubator/lucene.net/trunk/src/core/Search/MatchAllDocsQuery.cs
    incubator/lucene.net/trunk/src/core/Search/MultiPhraseQuery.cs
    incubator/lucene.net/trunk/src/core/Search/MultiSearcher.cs
    incubator/lucene.net/trunk/src/core/Search/MultiTermQuery.cs
    incubator/lucene.net/trunk/src/core/Search/ParallelMultiSearcher.cs
    incubator/lucene.net/trunk/src/core/Search/Payloads/PayloadNearQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Payloads/PayloadSpanUtil.cs
    incubator/lucene.net/trunk/src/core/Search/Payloads/PayloadTermQuery.cs
    incubator/lucene.net/trunk/src/core/Search/PhraseQuery.cs
    incubator/lucene.net/trunk/src/core/Search/PositiveScoresOnlyCollector.cs
    incubator/lucene.net/trunk/src/core/Search/Query.cs
    incubator/lucene.net/trunk/src/core/Search/QueryTermVector.cs
    incubator/lucene.net/trunk/src/core/Search/Similarity.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/NearSpansOrdered.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/NearSpansUnordered.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/SpanFirstQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/SpanNotQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/SpanOrQuery.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/SpanWeight.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/Spans.cs
    incubator/lucene.net/trunk/src/core/Search/Spans/TermSpans.cs
    incubator/lucene.net/trunk/src/core/Search/TermQuery.cs
    incubator/lucene.net/trunk/src/core/Search/TermScorer.cs
    incubator/lucene.net/trunk/src/core/Search/TimeLimitingCollector.cs
    incubator/lucene.net/trunk/src/core/Search/TopDocs.cs
    incubator/lucene.net/trunk/src/core/Search/TopFieldCollector.cs
    incubator/lucene.net/trunk/src/core/Search/TopScoreDocCollector.cs
    incubator/lucene.net/trunk/src/core/Search/Weight.cs
    incubator/lucene.net/trunk/src/core/Search/WildcardQuery.cs
    incubator/lucene.net/trunk/src/core/Store/BufferedIndexInput.cs
    incubator/lucene.net/trunk/src/core/Store/Directory.cs
    incubator/lucene.net/trunk/src/core/Store/FSDirectory.cs
    incubator/lucene.net/trunk/src/core/Util/AttributeSource.cs
    incubator/lucene.net/trunk/src/core/Util/FieldCacheSanityChecker.cs
    incubator/lucene.net/trunk/src/core/Util/OpenBitSet.cs
    incubator/lucene.net/trunk/src/core/Util/ReaderUtil.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Cn/TestChineseTokenizer.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/NumericPayloadTokenFilterTest.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilterTest.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilterTest.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/ShingleFilterTest.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/TestShingleMatrixFilter.cs
    incubator/lucene.net/trunk/test/contrib/Analyzers/Sinks/TokenTypeSinkTokenizerTest.cs
    incubator/lucene.net/trunk/test/contrib/Core/Index/SegmentsGenCommitTest.cs
    incubator/lucene.net/trunk/test/contrib/Core/Index/TermVectorEnumeratorTest.cs
    incubator/lucene.net/trunk/test/contrib/Core/Util/Cache/SegmentCacheTest.cs
    incubator/lucene.net/trunk/test/contrib/Snowball/Analysis/Snowball/TestSnowball.cs
    incubator/lucene.net/trunk/test/contrib/SpellChecker/Test/TestSpellChecker.cs
    incubator/lucene.net/trunk/test/core/Analysis/BaseTokenStreamTestCase.cs
    incubator/lucene.net/trunk/test/core/Analysis/TestKeywordAnalyzer.cs
    incubator/lucene.net/trunk/test/core/Analysis/TestNumericTokenStream.cs
    incubator/lucene.net/trunk/test/core/Analysis/TestToken.cs
    incubator/lucene.net/trunk/test/core/Analysis/Tokenattributes/TestSimpleAttributeImpls.cs
    incubator/lucene.net/trunk/test/core/Document/TestBinaryDocument.cs
    incubator/lucene.net/trunk/test/core/Index/TestAddIndexesNoOptimize.cs
    incubator/lucene.net/trunk/test/core/Index/TestAtomicUpdate.cs
    incubator/lucene.net/trunk/test/core/Index/TestBackwardsCompatibility.cs
    incubator/lucene.net/trunk/test/core/Index/TestConcurrentMergeScheduler.cs
    incubator/lucene.net/trunk/test/core/Index/TestCrash.cs
    incubator/lucene.net/trunk/test/core/Index/TestDeletionPolicy.cs
    incubator/lucene.net/trunk/test/core/Index/TestDirectoryReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestDoc.cs
    incubator/lucene.net/trunk/test/core/Index/TestDocumentWriter.cs
    incubator/lucene.net/trunk/test/core/Index/TestFieldsReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestFilterIndexReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexReaderClone.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexReaderCloneNorms.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexReaderReopen.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriter.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriterDelete.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriterExceptions.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriterMergePolicy.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriterMerging.cs
    incubator/lucene.net/trunk/test/core/Index/TestIndexWriterReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestIsCurrent.cs
    incubator/lucene.net/trunk/test/core/Index/TestNRTReaderWithThreads.cs
    incubator/lucene.net/trunk/test/core/Index/TestNorms.cs
    incubator/lucene.net/trunk/test/core/Index/TestOmitTf.cs
    incubator/lucene.net/trunk/test/core/Index/TestParallelReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestRollback.cs
    incubator/lucene.net/trunk/test/core/Index/TestSegmentMerger.cs
    incubator/lucene.net/trunk/test/core/Index/TestSegmentReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestStressIndexing2.cs
    incubator/lucene.net/trunk/test/core/Index/TestTermVectorsReader.cs
    incubator/lucene.net/trunk/test/core/Index/TestTermdocPerf.cs
    incubator/lucene.net/trunk/test/core/Index/TestThreadedOptimize.cs
    incubator/lucene.net/trunk/test/core/Index/TestTransactionRollback.cs
    incubator/lucene.net/trunk/test/core/Index/TestTransactions.cs
    incubator/lucene.net/trunk/test/core/QueryParser/TestMultiAnalyzer.cs
    incubator/lucene.net/trunk/test/core/QueryParser/TestQueryParser.cs
    incubator/lucene.net/trunk/test/core/Search/CheckHits.cs
    incubator/lucene.net/trunk/test/core/Search/Function/TestCustomScoreQuery.cs
    incubator/lucene.net/trunk/test/core/Search/Function/TestFieldScoreQuery.cs
    incubator/lucene.net/trunk/test/core/Search/Function/TestOrdValues.cs
    incubator/lucene.net/trunk/test/core/Search/JustCompileSearch.cs
    incubator/lucene.net/trunk/test/core/Search/QueryUtils.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/JustCompileSearchSpans.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/TestBasics.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/TestNearSpansOrdered.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/TestPayloadSpans.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/TestSpansAdvanced.cs
    incubator/lucene.net/trunk/test/core/Search/Spans/TestSpansAdvanced2.cs
    incubator/lucene.net/trunk/test/core/Search/TestBoolean2.cs
    incubator/lucene.net/trunk/test/core/Search/TestBooleanQuery.cs
    incubator/lucene.net/trunk/test/core/Search/TestDocBoost.cs
    incubator/lucene.net/trunk/test/core/Search/TestElevationComparator.cs
    incubator/lucene.net/trunk/test/core/Search/TestFieldCacheRangeFilter.cs
    incubator/lucene.net/trunk/test/core/Search/TestFieldCacheTermsFilter.cs
    incubator/lucene.net/trunk/test/core/Search/TestFuzzyQuery.cs
    incubator/lucene.net/trunk/test/core/Search/TestMultiTermConstantScore.cs
    incubator/lucene.net/trunk/test/core/Search/TestMultiThreadTermVectors.cs
    incubator/lucene.net/trunk/test/core/Search/TestNumericRangeQuery32.cs
    incubator/lucene.net/trunk/test/core/Search/TestNumericRangeQuery64.cs
    incubator/lucene.net/trunk/test/core/Search/TestPositionIncrement.cs
    incubator/lucene.net/trunk/test/core/Search/TestPositiveScoresOnlyCollector.cs
    incubator/lucene.net/trunk/test/core/Search/TestScoreCachingWrappingScorer.cs
    incubator/lucene.net/trunk/test/core/Search/TestScorerPerf.cs
    incubator/lucene.net/trunk/test/core/Search/TestSetNorm.cs
    incubator/lucene.net/trunk/test/core/Search/TestSimilarity.cs
    incubator/lucene.net/trunk/test/core/Search/TestSort.cs
    incubator/lucene.net/trunk/test/core/Search/TestTermRangeFilter.cs
    incubator/lucene.net/trunk/test/core/Search/TestTermRangeQuery.cs
    incubator/lucene.net/trunk/test/core/Search/TestTermScorer.cs
    incubator/lucene.net/trunk/test/core/Search/TestTermVectors.cs
    incubator/lucene.net/trunk/test/core/Search/TestTimeLimitingCollector.cs
    incubator/lucene.net/trunk/test/core/Search/TestTopDocsCollector.cs
    incubator/lucene.net/trunk/test/core/Search/TestTopScoreDocCollector.cs
    incubator/lucene.net/trunk/test/core/Store/TestRAMDirectory.cs
    incubator/lucene.net/trunk/test/core/Support/TestOldPatches.cs
    incubator/lucene.net/trunk/test/core/Util/TestAttributeSource.cs

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/AR/ArabicAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/AR/ArabicAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/AR/ArabicAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/AR/ArabicAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -183,7 +183,7 @@ namespace Lucene.Net.Analysis.AR
          */
         public override TokenStream ReusableTokenStream(string fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.AR
                                                 streams.Result, stoptable);
                 streams.Result = new ArabicNormalizationFilter(streams.Result);
                 streams.Result = new ArabicStemFilter(streams.Result);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/BR/BrazilianAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/BR/BrazilianAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/BR/BrazilianAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/BR/BrazilianAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -167,7 +167,7 @@ namespace Lucene.Net.Analysis.BR
         public void SetStemExclusionTable(params string[] exclusionlist)
         {
             excltable = StopFilter.MakeStopSet(exclusionlist);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -178,7 +178,7 @@ namespace Lucene.Net.Analysis.BR
         public void SetStemExclusionTable(IDictionary<string, string> exclusionlist)
         {
             excltable = new HashSet<string>(exclusionlist.Keys);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -189,7 +189,7 @@ namespace Lucene.Net.Analysis.BR
         public void SetStemExclusionTable(FileInfo exclusionlist)
         {
             excltable = WordlistLoader.GetWordSet(exclusionlist);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -227,7 +227,7 @@ namespace Lucene.Net.Analysis.BR
 
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams) GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams) PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -237,7 +237,7 @@ namespace Lucene.Net.Analysis.BR
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, stoptable);
                 streams.result = new BrazilianStemFilter(streams.result, excltable);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -135,14 +135,14 @@ namespace Lucene.Net.Analysis.CJK
         public override sealed TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
             /* tokenStream() is final, no back compat issue */
-            SavedStreams streams = (SavedStreams) GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams) PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
                 streams.source = new CJKTokenizer(reader);
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.source, stopTable);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKTokenizer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKTokenizer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKTokenizer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/CJK/CJKTokenizer.cs Wed Mar 21 06:04:26 2012
@@ -361,7 +361,7 @@ namespace Lucene.Net.Analysis.CJK
                 {
                     termAtt.SetTermBuffer(buffer, 0, length);
                     offsetAtt.SetOffset(CorrectOffset(start), CorrectOffset(start + length));
-                    typeAtt.SetType(TOKEN_TYPE_NAMES[tokenType]);
+                    typeAtt.Type = TOKEN_TYPE_NAMES[tokenType];
                     return true;
                 }
                 else if (dataLen == 0)

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Cn/ChineseAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Cn/ChineseAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Cn/ChineseAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Cn/ChineseAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -67,13 +67,13 @@ namespace Lucene.Net.Analysis.Cn
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
             /* tokenStream() is final, no back compat issue */
-            SavedStreams streams = (SavedStreams) GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams) PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
                 streams.source = new ChineseTokenizer(reader);
                 streams.result = new ChineseFilter(streams.source);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Compound/CompoundWordTokenFilterBase.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Compound/CompoundWordTokenFilterBase.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Compound/CompoundWordTokenFilterBase.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Compound/CompoundWordTokenFilterBase.cs Wed Mar 21 06:04:26 2012
@@ -136,8 +136,8 @@ namespace Lucene.Net.Analysis.Compound
             ClearAttributes();
             termAtt.SetTermBuffer(token.TermBuffer(), 0, token.TermLength());
             flagsAtt.Flags = token.Flags;
-            typeAtt.SetType(token.Type());
-            offsetAtt.SetOffset(token.StartOffset(), token.EndOffset());
+            typeAtt.Type = token.Type;
+            offsetAtt.SetOffset(token.StartOffset, token.EndOffset);
             posIncAtt.PositionIncrement = token.PositionIncrement;
             payloadAtt.Payload = token.Payload;
         }
@@ -155,10 +155,10 @@ namespace Lucene.Net.Analysis.Compound
                 return false;
 
             wrapper.SetTermBuffer(termAtt.TermBuffer(), 0, termAtt.TermLength());
-            wrapper.SetStartOffset(offsetAtt.StartOffset());
-            wrapper.SetEndOffset(offsetAtt.EndOffset());
+            wrapper.StartOffset = offsetAtt.StartOffset;
+            wrapper.EndOffset = offsetAtt.EndOffset;
             wrapper.Flags = flagsAtt.Flags;
-            wrapper.SetType(typeAtt.Type());
+            wrapper.Type = typeAtt.Type;
             wrapper.PositionIncrement = posIncAtt.PositionIncrement;
             wrapper.Payload = payloadAtt.Payload;
 
@@ -200,7 +200,7 @@ namespace Lucene.Net.Analysis.Compound
         protected Token CreateToken(int offset, int length,
             Token prototype)
         {
-            int newStart = prototype.StartOffset() + offset;
+            int newStart = prototype.StartOffset + offset;
             Token t = prototype.Clone(prototype.TermBuffer(), offset, length, newStart, newStart + length);
             t.PositionIncrement = 0;
             return t;

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Cz/CzechAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Cz/CzechAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Cz/CzechAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Cz/CzechAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -149,7 +149,7 @@ public sealed class CzechAnalyzer : Anal
      *             and {@link #CzechAnalyzer(Version, Set)} instead
      */
     public void LoadStopWords( Stream wordfile, System.Text.Encoding encoding ) {
-        SetPreviousTokenStream(null); // force a new stopfilter to be created
+        PreviousTokenStream = null; // force a new stopfilter to be created
         if ( wordfile == null )
         {
             stoptable = new HashSet<string>();
@@ -202,7 +202,7 @@ public sealed class CzechAnalyzer : Anal
      */
 	public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
     {
-      SavedStreams streams = (SavedStreams) GetPreviousTokenStream();
+      SavedStreams streams = (SavedStreams) PreviousTokenStream;
       if (streams == null) {
         streams = new SavedStreams();
         streams.source = new StandardTokenizer(matchVersion, reader);
@@ -210,7 +210,7 @@ public sealed class CzechAnalyzer : Anal
         streams.result = new LowerCaseFilter(streams.result);
         streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                         streams.result, stoptable);
-        SetPreviousTokenStream(streams);
+        PreviousTokenStream = streams;
       } else {
         streams.source.Reset(reader);
       }

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/De/GermanAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/De/GermanAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/De/GermanAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/De/GermanAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -1,4 +1,4 @@
-/*
+/*
  *
  * Licensed to the Apache Software Foundation (ASF) under one
  * or more contributor license agreements.  See the NOTICE file
@@ -167,7 +167,7 @@ namespace Lucene.Net.Analysis.De
         public void SetStemExclusionTable(String[] exclusionlist)
         {
             exclusionSet = StopFilter.MakeStopSet(exclusionlist);
-            SetPreviousTokenStream(null);
+            PreviousTokenStream = null;
         }
 
         /// <summary>
@@ -177,7 +177,7 @@ namespace Lucene.Net.Analysis.De
         public void SetStemExclusionTable(IDictionary<string, string> exclusionlist)
         {
             exclusionSet = new HashSet<string>(exclusionlist.Keys);
-            SetPreviousTokenStream(null);
+            PreviousTokenStream = null;
         }
 
         /// <summary>
@@ -187,7 +187,7 @@ namespace Lucene.Net.Analysis.De
         public void SetStemExclusionTable(FileInfo exclusionlist)
         {
             exclusionSet = WordlistLoader.GetWordSet(exclusionlist);
-            SetPreviousTokenStream(null);
+            PreviousTokenStream = null;
         }
 
         /// <summary>

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/El/GreekAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/El/GreekAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/El/GreekAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/El/GreekAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -1,4 +1,4 @@
-/*
+/*
  *
  * Licensed to the Apache Software Foundation (ASF) under one
  * or more contributor license agreements.  See the NOTICE file
@@ -156,7 +156,7 @@ namespace Lucene.Net.Analyzers.El
          */
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -164,7 +164,7 @@ namespace Lucene.Net.Analyzers.El
                 streams.result = new GreekLowerCaseFilter(streams.source);
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, stopSet);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Fa/PersianAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Fa/PersianAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Fa/PersianAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Fa/PersianAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -209,7 +209,7 @@ namespace Lucene.Net.Analyzers.Fa
          */
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -224,7 +224,7 @@ namespace Lucene.Net.Analyzers.Fa
                  */
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, stoptable);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Fr/FrenchAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Fr/FrenchAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Fr/FrenchAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Fr/FrenchAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -1,4 +1,4 @@
-/*
+/*
  *
  * Licensed to the Apache Software Foundation (ASF) under one
  * or more contributor license agreements.  See the NOTICE file
@@ -179,7 +179,7 @@ namespace Lucene.Net.Analysis.Fr
         public void SetStemExclusionTable(params string[] exclusionlist)
         {
             excltable = StopFilter.MakeStopSet(exclusionlist);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -189,7 +189,7 @@ namespace Lucene.Net.Analysis.Fr
         public void SetStemExclusionTable(IDictionary<string, string> exclusionlist)
         {
             excltable = new HashSet<string>(exclusionlist.Keys);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -200,7 +200,7 @@ namespace Lucene.Net.Analysis.Fr
         public void SetStemExclusionTable(FileInfo exclusionlist)
         {
             excltable = new HashSet<string>(WordlistLoader.GetWordSet(exclusionlist));
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -239,7 +239,7 @@ namespace Lucene.Net.Analysis.Fr
          */
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -250,7 +250,7 @@ namespace Lucene.Net.Analysis.Fr
                 streams.result = new FrenchStemFilter(streams.result, excltable);
                 // Convert to lowercase after stemming!
                 streams.result = new LowerCaseFilter(streams.result);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAndSuffixAwareTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAndSuffixAwareTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAndSuffixAwareTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAndSuffixAwareTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -46,15 +46,15 @@ namespace Lucene.Net.Analyzers.Miscellan
 
         public Token UpdateInputToken(Token inputToken, Token lastPrefixToken)
         {
-            inputToken.SetStartOffset(lastPrefixToken.EndOffset() + inputToken.StartOffset());
-            inputToken.SetEndOffset(lastPrefixToken.EndOffset() + inputToken.EndOffset());
+            inputToken.StartOffset = lastPrefixToken.EndOffset + inputToken.StartOffset;
+            inputToken.EndOffset = lastPrefixToken.EndOffset + inputToken.EndOffset;
             return inputToken;
         }
 
         public Token UpdateSuffixToken(Token suffixToken, Token lastInputToken)
         {
-            suffixToken.SetStartOffset(lastInputToken.EndOffset() + suffixToken.StartOffset());
-            suffixToken.SetEndOffset(lastInputToken.EndOffset() + suffixToken.EndOffset());
+            suffixToken.StartOffset = lastInputToken.EndOffset + suffixToken.StartOffset;
+            suffixToken.EndOffset = lastInputToken.EndOffset + suffixToken.EndOffset;
             return suffixToken;
         }
 

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAwareTokenStream.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAwareTokenStream.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAwareTokenStream.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Miscellaneous/PrefixAwareTokenStream.cs Wed Mar 21 06:04:26 2012
@@ -121,8 +121,8 @@ namespace Lucene.Net.Analyzers.Miscellan
             _termAtt.SetTermBuffer(token.TermBuffer(), 0, token.TermLength());
             _posIncrAtt.PositionIncrement = token.PositionIncrement;
             _flagsAtt.Flags =token.Flags;
-            _offsetAtt.SetOffset(token.StartOffset(), token.EndOffset());
-            _typeAtt.SetType(token.Type());
+            _offsetAtt.SetOffset(token.StartOffset, token.EndOffset);
+            _typeAtt.Type = token.Type;
             _payloadAtt.Payload = token.Payload;
         }
 
@@ -132,8 +132,8 @@ namespace Lucene.Net.Analyzers.Miscellan
             token.SetTermBuffer(_pTermAtt.TermBuffer(), 0, _pTermAtt.TermLength());
             token.PositionIncrement = _pPosIncrAtt.PositionIncrement;
             token.Flags = _pFlagsAtt.Flags;
-            token.SetOffset(_pOffsetAtt.StartOffset(), _pOffsetAtt.EndOffset());
-            token.SetType(_pTypeAtt.Type());
+            token.SetOffset(_pOffsetAtt.StartOffset, _pOffsetAtt.EndOffset);
+            token.Type = _pTypeAtt.Type;
             token.Payload = _pPayloadAtt.Payload;
             return token;
         }
@@ -144,8 +144,8 @@ namespace Lucene.Net.Analyzers.Miscellan
             token.SetTermBuffer(_termAtt.TermBuffer(), 0, _termAtt.TermLength());
             token.PositionIncrement = _posIncrAtt.PositionIncrement;
             token.Flags = _flagsAtt.Flags;
-            token.SetOffset(_offsetAtt.StartOffset(), _offsetAtt.EndOffset());
-            token.SetType(_typeAtt.Type());
+            token.SetOffset(_offsetAtt.StartOffset, _offsetAtt.EndOffset);
+            token.Type = _typeAtt.Type;
             token.Payload = _payloadAtt.Payload;
             return token;
         }
@@ -158,8 +158,8 @@ namespace Lucene.Net.Analyzers.Miscellan
         /// <returns>consumer token</returns>
         public virtual Token UpdateSuffixToken(Token suffixToken, Token lastPrefixToken)
         {
-            suffixToken.SetStartOffset(lastPrefixToken.EndOffset() + suffixToken.StartOffset());
-            suffixToken.SetEndOffset(lastPrefixToken.EndOffset() + suffixToken.EndOffset());
+            suffixToken.StartOffset = lastPrefixToken.EndOffset + suffixToken.StartOffset;
+            suffixToken.EndOffset = lastPrefixToken.EndOffset + suffixToken.EndOffset;
             return suffixToken;
         }
 

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/EdgeNGramTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/EdgeNGramTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/EdgeNGramTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/EdgeNGramTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -158,7 +158,7 @@ namespace Lucene.Net.Analysis.NGram
                         curTermBuffer = (char[])termAtt.TermBuffer().Clone();
                         curTermLength = termAtt.TermLength();
                         curGramSize = minGram;
-                        tokStart = offsetAtt.StartOffset();
+                        tokStart = offsetAtt.StartOffset;
                     }
                 }
                 if (curGramSize <= maxGram)

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/NGramTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/NGramTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/NGramTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/NGram/NGramTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis.NGram
                         curTermLength = termAtt.TermLength();
                         curGramSize = minGram;
                         curPos = 0;
-                        tokStart = offsetAtt.StartOffset();
+                        tokStart = offsetAtt.StartOffset;
                     }
                 }
                 while (curGramSize <= maxGram)

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Nl/DutchAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Nl/DutchAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Nl/DutchAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Nl/DutchAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -175,7 +175,7 @@ namespace Lucene.Net.Analysis.Nl
         public void SetStemExclusionTable(params string[] exclusionlist)
         {
             excltable = StopFilter.MakeStopSet(exclusionlist);
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -185,7 +185,7 @@ namespace Lucene.Net.Analysis.Nl
         public void SetStemExclusionTable(HashSet<string> exclusionlist)
         {
             excltable = exclusionlist;
-            SetPreviousTokenStream(null); // force a new stemmer to be created
+            PreviousTokenStream = null; // force a new stemmer to be created
         }
 
         /**
@@ -197,7 +197,7 @@ namespace Lucene.Net.Analysis.Nl
             try
             {
                 excltable = WordlistLoader.GetWordSet(exclusionlist);
-                SetPreviousTokenStream(null); // force a new stemmer to be created
+                PreviousTokenStream = null; // force a new stemmer to be created
             }
             catch (IOException e)
             {
@@ -216,7 +216,7 @@ namespace Lucene.Net.Analysis.Nl
             try
             {
                 stemdict = WordlistLoader.GetStemDict(stemdictFile);
-                SetPreviousTokenStream(null); // force a new stemmer to be created
+                PreviousTokenStream = null; // force a new stemmer to be created
             }
             catch (IOException e)
             {
@@ -267,7 +267,7 @@ namespace Lucene.Net.Analysis.Nl
                 return TokenStream(fieldName, reader);
             }
 
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -276,7 +276,7 @@ namespace Lucene.Net.Analysis.Nl
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, stoptable);
                 streams.result = new DutchStemFilter(streams.result, excltable, stemdict);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/NumericPayloadTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/NumericPayloadTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/NumericPayloadTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/NumericPayloadTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -54,7 +54,7 @@ namespace Lucene.Net.Analyzers.Payloads
         {
             if (input.IncrementToken())
             {
-                if (typeAtt.Type().Equals(typeMatch))
+                if (typeAtt.Type.Equals(typeMatch))
                     payloadAtt.Payload = thePayload;
                 return true;
             }

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -51,8 +51,8 @@ namespace Lucene.Net.Analyzers.Payloads
             if (input.IncrementToken())
             {
                 byte[] data = new byte[8];
-                PayloadHelper.EncodeInt(offsetAtt.StartOffset(), data, 0);
-                PayloadHelper.EncodeInt(offsetAtt.EndOffset(), data, 4);
+                PayloadHelper.EncodeInt(offsetAtt.StartOffset, data, 0);
+                PayloadHelper.EncodeInt(offsetAtt.EndOffset, data, 4);
                 Payload payload = new Payload(data);
                 payAtt.Payload = payload;
                 return true;

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilter.cs Wed Mar 21 06:04:26 2012
@@ -49,7 +49,7 @@ namespace Lucene.Net.Analyzers.Payloads
         {
             if (input.IncrementToken())
             {
-                String type = typeAtt.Type();
+                String type = typeAtt.Type;
                 if (type != null && type.Equals("") == false)
                 {
                     payloadAtt.Payload = new Payload(Encoding.UTF8.GetBytes(type));

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Query/QueryAutoStopWordAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Query/QueryAutoStopWordAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Query/QueryAutoStopWordAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Query/QueryAutoStopWordAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -133,7 +133,7 @@ public class QueryAutoStopWordAnalyzer :
    */
   public int AddStopWords(IndexReader reader, String fieldName, float maxPercentDocs) 
   {
-    return AddStopWords(reader, fieldName, (int) (reader.NumDocs * maxPercentDocs));
+    return AddStopWords(reader, fieldName, (int) (reader.GetNumDocs() * maxPercentDocs));
   }
 
   /**
@@ -170,7 +170,7 @@ public class QueryAutoStopWordAnalyzer :
     /* if the stopwords for a field are changed,
      * then saved streams for that field are erased.
      */
-    IDictionary<String,SavedStreams> streamMap = (IDictionary<String,SavedStreams>) GetPreviousTokenStream();
+    IDictionary<String,SavedStreams> streamMap = (IDictionary<String,SavedStreams>) PreviousTokenStream;
     if (streamMap != null)
       streamMap.Remove(fieldName);
     
@@ -213,10 +213,10 @@ public class QueryAutoStopWordAnalyzer :
     }
 
     /* map of SavedStreams for each field */
-    IDictionary<String, SavedStreams> streamMap = (IDictionary<String, SavedStreams>)GetPreviousTokenStream();
+    IDictionary<String, SavedStreams> streamMap = (IDictionary<String, SavedStreams>)PreviousTokenStream;
     if (streamMap == null) {
       streamMap = new HashMap<String, SavedStreams>();
-      SetPreviousTokenStream(streamMap);
+      PreviousTokenStream = streamMap;
     }
 
     SavedStreams streams = streamMap[fieldName];

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Ru/RussianAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Ru/RussianAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Ru/RussianAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Ru/RussianAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -151,7 +151,7 @@ namespace Lucene.Net.Analysis.Ru
          */
         public override TokenStream ReusableTokenStream(String fieldName, TextReader reader)
         {
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -160,7 +160,7 @@ namespace Lucene.Net.Analysis.Ru
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, stopSet);
                 streams.result = new RussianStemFilter(streams.result);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleAnalyzerWrapper.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleAnalyzerWrapper.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleAnalyzerWrapper.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleAnalyzerWrapper.cs Wed Mar 21 06:04:26 2012
@@ -137,13 +137,13 @@ namespace Lucene.Net.Analyzers.Shingle
                 return TokenStream(fieldName, reader);
             }
 
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
                 streams.wrapped = defaultAnalyzer.ReusableTokenStream(fieldName, reader);
                 streams.shingle = new ShingleFilter(streams.wrapped);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleFilter.cs Wed Mar 21 06:04:26 2012
@@ -201,8 +201,8 @@ namespace Lucene.Net.Analyzers.Shingle
                 if (shingleBufferPosition < shingleBuf.Count)
                 {
                     RestoreState(nextToken);
-                    typeAtt.SetType(tokenType);
-                    offsetAtt.SetOffset(offsetAtt.StartOffset(), endOffsets[shingleBufferPosition]);
+                    typeAtt.Type = tokenType;
+                    offsetAtt.SetOffset(offsetAtt.StartOffset, endOffsets[shingleBufferPosition]);
                     StringBuilder buf = shingles[shingleBufferPosition];
                     int termLength = buf.Length;
                     char[] TermBuffer = termAtt.TermBuffer();
@@ -269,7 +269,7 @@ namespace Lucene.Net.Analyzers.Shingle
                     }
                     numFillerTokensToInsert--;
                     // A filler token occupies no space
-                    offsetAtt.SetOffset(offsetAtt.StartOffset(), offsetAtt.StartOffset());
+                    offsetAtt.SetOffset(offsetAtt.StartOffset, offsetAtt.StartOffset);
                     termAtt.SetTermBuffer(FILLER_TOKEN, 0, FILLER_TOKEN.Length);
                     return true;
                 }
@@ -361,7 +361,7 @@ namespace Lucene.Net.Analyzers.Shingle
                     shingles[j].Append(termAtt.TermBuffer().Take(termAtt.TermLength()).ToArray());
                 }
 
-                endOffsets[i] = offsetAtt.EndOffset();
+                endOffsets[i] = offsetAtt.EndOffset;
                 i++;
             }
 

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleMatrixFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleMatrixFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleMatrixFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Shingle/ShingleMatrixFilter.cs Wed Mar 21 06:04:26 2012
@@ -319,8 +319,8 @@ namespace Lucene.Net.Analyzers.Shingle
             _termAtt.SetTermBuffer(token.TermBuffer(), 0, token.TermLength());
             _posIncrAtt.PositionIncrement = token.PositionIncrement;
             _flagsAtt.Flags = token.Flags;
-            _offsetAtt.SetOffset(token.StartOffset(), token.EndOffset());
-            _typeAtt.SetType(token.Type());
+            _offsetAtt.SetOffset(token.StartOffset, token.EndOffset);
+            _typeAtt.Type = token.Type;
             _payloadAtt.Payload = token.Payload;
 
             return true;
@@ -333,8 +333,8 @@ namespace Lucene.Net.Analyzers.Shingle
             token.SetTermBuffer(_inTermAtt.TermBuffer(), 0, _inTermAtt.TermLength());
             token.PositionIncrement = _inPosIncrAtt.PositionIncrement;
             token.Flags = _inFlagsAtt.Flags;
-            token.SetOffset(_inOffsetAtt.StartOffset(), _inOffsetAtt.EndOffset());
-            token.SetType(_inTypeAtt.Type());
+            token.SetOffset(_inOffsetAtt.StartOffset, _inOffsetAtt.EndOffset);
+            token.Type = _inTypeAtt.Type;
             token.Payload = _inPayloadAtt.Payload;
             return token;
         }
@@ -345,8 +345,8 @@ namespace Lucene.Net.Analyzers.Shingle
             token.SetTermBuffer(_termAtt.TermBuffer(), 0, _termAtt.TermLength());
             token.PositionIncrement = _posIncrAtt.PositionIncrement;
             token.Flags = _flagsAtt.Flags;
-            token.SetOffset(_offsetAtt.StartOffset(), _offsetAtt.EndOffset());
-            token.SetType(_typeAtt.Type());
+            token.SetOffset(_offsetAtt.StartOffset, _offsetAtt.EndOffset);
+            token.Type = _typeAtt.Type;
             token.Payload = _payloadAtt.Payload;
             return token;
         }
@@ -529,11 +529,11 @@ namespace Lucene.Net.Analyzers.Shingle
         /// <param name="currentPermuationTokens">tokens of the current permutation of rows in the matrix. </param>
         public void UpdateToken(Token token, List<Token> shingle, int currentPermutationStartOffset, List<Row> currentPermutationRows, List<Token> currentPermuationTokens)
         {
-            token.SetType(typeof(ShingleMatrixFilter).Name);
+            token.Type = typeof(ShingleMatrixFilter).Name;
             token.Flags = 0;
             token.PositionIncrement = 1;
-            token.SetStartOffset((shingle[0]).StartOffset());
-            token.SetEndOffset(shingle[shingle.Count - 1].EndOffset());
+            token.StartOffset = (shingle[0]).StartOffset;
+            token.EndOffset = shingle[shingle.Count - 1].EndOffset;
 
             _settingsCodec.SetWeight(
                 token, 

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Sinks/TokenTypeSinkFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Sinks/TokenTypeSinkFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Sinks/TokenTypeSinkFilter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Sinks/TokenTypeSinkFilter.cs Wed Mar 21 06:04:26 2012
@@ -45,7 +45,7 @@ namespace Lucene.Net.Analysis.Sinks
                 typeAtt = source.AddAttribute<TypeAttribute>();
             }
 
-            return typeToMatch.Equals(typeAtt.Type());
+            return typeToMatch.Equals(typeAtt.Type);
         }
     }
 }

Modified: incubator/lucene.net/trunk/src/contrib/Analyzers/Th/ThaiAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Analyzers/Th/ThaiAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Analyzers/Th/ThaiAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Analyzers/Th/ThaiAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -72,7 +72,7 @@ namespace Lucene.Net.Analysis.Th
                 return TokenStream(fieldName, reader);
             }
 
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -81,7 +81,7 @@ namespace Lucene.Net.Analysis.Th
                 streams.result = new ThaiWordFilter(streams.result);
                 streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                 streams.result, StopAnalyzer.ENGLISH_STOP_WORDS_SET);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/Core/Index/TermVectorEnumerator.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Core/Index/TermVectorEnumerator.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Core/Index/TermVectorEnumerator.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Core/Index/TermVectorEnumerator.cs Wed Mar 21 06:04:26 2012
@@ -27,7 +27,7 @@ namespace Lucene.Net.Index
     /// Class to allow for enumerating over the documents in the index to 
     /// retrieve the term vector for each one.
     /// </summary>
-    public class TermVectorEnumerator : IEnumerator<TermFreqVector>, IEnumerable<TermFreqVector>
+    public class TermVectorEnumerator : IEnumerator<ITermFreqVector>, IEnumerable<ITermFreqVector>
     {
         /// <summary>
         /// Current document being accessed.
@@ -63,7 +63,7 @@ namespace Lucene.Net.Index
 
         #region IEnumerator<TermFreqVector> Members
 
-        public TermFreqVector Current
+        public ITermFreqVector Current
         {
             get { return this.CurrentVector(); }
         }
@@ -101,9 +101,9 @@ namespace Lucene.Net.Index
 
         #region IEnumerable<TermFreqVector> Members
 
-        public IEnumerator<TermFreqVector> GetEnumerator()
+        public IEnumerator<ITermFreqVector> GetEnumerator()
         {
-            return (IEnumerator<TermFreqVector>)this;
+            return (IEnumerator<ITermFreqVector>)this;
         }
 
         #endregion
@@ -112,7 +112,7 @@ namespace Lucene.Net.Index
 
         IEnumerator IEnumerable.GetEnumerator()
         {
-            return (IEnumerator<TermFreqVector>)this;
+            return (IEnumerator<ITermFreqVector>)this;
         }
 
         #endregion
@@ -121,7 +121,7 @@ namespace Lucene.Net.Index
         /// Retrieve the current TermFreqVector from the index.
         /// </summary>
         /// <returns>The current TermFreqVector.</returns>
-        private TermFreqVector CurrentVector()
+        private ITermFreqVector CurrentVector()
         {
             if (this.reader.IsDeleted(this.document))
             {
@@ -129,7 +129,7 @@ namespace Lucene.Net.Index
             }
             else
             {
-                TermFreqVector vector = this.reader.GetTermFreqVector(this.document, this.fieldName);
+                ITermFreqVector vector = this.reader.GetTermFreqVector(this.document, this.fieldName);
                 if (vector == null)
                 {
                     vector = this.emptyVector;
@@ -144,7 +144,7 @@ namespace Lucene.Net.Index
     /// with a deleted document or a document that does not have the field
     /// that is being enumerated.
     /// </summary>
-    public class EmptyVector : TermFreqVector
+    public class EmptyVector : ITermFreqVector
     {
         private string field;
 
@@ -159,14 +159,14 @@ namespace Lucene.Net.Index
 
         #region TermFreqVector Members
 
-        public string GetField()
+        public string Field
         {
-            return this.field;
+            get { return this.field; }
         }
 
-        public int Size()
+        public int Size
         {
-            return 0;
+            get { return 0; }
         }
 
         public string[] GetTerms()

Modified: incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/FieldTermStack.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/FieldTermStack.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/FieldTermStack.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/FieldTermStack.cs Wed Mar 21 06:04:26 2012
@@ -80,7 +80,7 @@ namespace Lucene.Net.Search.Vectorhighli
             VectorHighlightMapper tfv = new VectorHighlightMapper(termSet);    
             reader.GetTermFreqVector(docId, fieldName, tfv);
             
-            if (tfv.Size()==0) return; // just return to make null snippets
+            if (tfv.Size==0) return; // just return to make null snippets
             
             string[] terms = tfv.GetTerms();
             foreach (String term in terms)

Modified: incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/VectorHighlightMapper.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/VectorHighlightMapper.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/VectorHighlightMapper.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/FastVectorHighlighter/VectorHighlightMapper.cs Wed Mar 21 06:04:26 2012
@@ -23,7 +23,7 @@ using Lucene.Net.Index;
 
 namespace Lucene.Net.Search.Vectorhighlight
 {
-    public class VectorHighlightMapper : TermVectorMapper, TermFreqVector, TermPositionVector
+    public class VectorHighlightMapper : TermVectorMapper, ITermFreqVector, TermPositionVector
     {
         private readonly List<string> _terms;
         private Dictionary<string, TermVectorOffsetInfo[]> _tvoi;
@@ -69,14 +69,14 @@ namespace Lucene.Net.Search.Vectorhighli
             }
         }
 
-        public string GetField()
+        public string Field
         {
-            return _field;
+            get { return _field; }
         }
 
-        public int Size()
+        public int Size
         {
-            return _tvoi.Count;
+            get { return _tvoi.Count; }
         }
 
         public string[] GetTerms()

Modified: incubator/lucene.net/trunk/src/contrib/Highlighter/QueryTermExtractor.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Highlighter/QueryTermExtractor.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Highlighter/QueryTermExtractor.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Highlighter/QueryTermExtractor.cs Wed Mar 21 06:04:26 2012
@@ -63,7 +63,7 @@ namespace Lucene.Net.Highlight
 		public static WeightedTerm[] GetIdfWeightedTerms(Query query, IndexReader reader, System.String fieldName)
 		{
 			WeightedTerm[] terms = GetTerms(query, false, fieldName);
-			int totalNumDocs = reader.NumDocs;
+			int totalNumDocs = reader.GetNumDocs();
 			for (int i = 0; i < terms.Length; i++)
 			{
 				try

Modified: incubator/lucene.net/trunk/src/contrib/Highlighter/SimpleFragmenter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Highlighter/SimpleFragmenter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Highlighter/SimpleFragmenter.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Highlighter/SimpleFragmenter.cs Wed Mar 21 06:04:26 2012
@@ -59,7 +59,7 @@ namespace Lucene.Net.Highlight
 		*/
 		public virtual bool IsNewFragment(Token token)
 		{
-			bool isNewFrag = token.EndOffset() >= (fragmentSize * currentNumFrags);
+			bool isNewFrag = token.EndOffset >= (fragmentSize * currentNumFrags);
 			if (isNewFrag)
 			{
 				currentNumFrags++;

Modified: incubator/lucene.net/trunk/src/contrib/Highlighter/TokenGroup.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Highlighter/TokenGroup.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Highlighter/TokenGroup.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Highlighter/TokenGroup.cs Wed Mar 21 06:04:26 2012
@@ -46,25 +46,25 @@ namespace Lucene.Net.Highlight
 			{
 				if (numTokens == 0)
 				{
-					startOffset = matchStartOffset = token.StartOffset();
-					endOffset = matchEndOffset = token.EndOffset();
+					startOffset = matchStartOffset = token.StartOffset;
+					endOffset = matchEndOffset = token.EndOffset;
 					tot += score;
 				}
 				else
 				{
-					startOffset = Math.Min(startOffset, token.StartOffset());
-					endOffset = Math.Max(endOffset, token.EndOffset());
+					startOffset = Math.Min(startOffset, token.StartOffset);
+					endOffset = Math.Max(endOffset, token.EndOffset);
 					if (score > 0)
 					{
 						if (tot == 0)
 						{
-							matchStartOffset = token.StartOffset();
-							matchEndOffset = token.EndOffset();
+							matchStartOffset = token.StartOffset;
+							matchEndOffset = token.EndOffset;
 						}
 						else
 						{
-							matchStartOffset = Math.Min(matchStartOffset, token.StartOffset());
-							matchEndOffset = Math.Max(matchEndOffset, token.EndOffset());
+							matchStartOffset = Math.Min(matchStartOffset, token.StartOffset);
+							matchEndOffset = Math.Max(matchEndOffset, token.EndOffset);
 						}
 						tot += score;
 					}
@@ -77,7 +77,7 @@ namespace Lucene.Net.Highlight
 		
 		internal virtual bool IsDistinct(Token token)
 		{
-			return token.StartOffset() >= endOffset;
+			return token.StartOffset >= endOffset;
 		}
 		
 		

Modified: incubator/lucene.net/trunk/src/contrib/Highlighter/TokenSources.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Highlighter/TokenSources.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Highlighter/TokenSources.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Highlighter/TokenSources.cs Wed Mar 21 06:04:26 2012
@@ -20,12 +20,12 @@
 */
 using System;
 using Lucene.Net.Analysis.Tokenattributes;
+using Lucene.Net.Index;
 using Analyzer = Lucene.Net.Analysis.Analyzer;
 using Token = Lucene.Net.Analysis.Token;
 using TokenStream = Lucene.Net.Analysis.TokenStream;
 using Document = Lucene.Net.Documents.Document;
 using IndexReader = Lucene.Net.Index.IndexReader;
-using TermFreqVector = Lucene.Net.Index.TermFreqVector;
 using TermPositionVector = Lucene.Net.Index.TermPositionVector;
 using TermVectorOffsetInfo = Lucene.Net.Index.TermVectorOffsetInfo;
 
@@ -62,7 +62,7 @@ namespace Lucene.Net.Highlight
                 ClearAttributes();
                 Token token = tokens[currentToken++];
                 termAtt.SetTermBuffer(token.Term());
-                offsetAtt.SetOffset(token.StartOffset(), token.EndOffset());
+                offsetAtt.SetOffset(token.StartOffset, token.EndOffset);
                 return true;
 			}
 
@@ -77,9 +77,9 @@ namespace Lucene.Net.Highlight
 			{
 				Token t1 = (Token) o1;
 				Token t2 = (Token) o2;
-				if (t1.StartOffset() > t2.StartOffset())
+				if (t1.StartOffset > t2.StartOffset)
 					return 1;
-				if (t1.StartOffset() < t2.StartOffset())
+				if (t1.StartOffset < t2.StartOffset)
 					return - 1;
 				return 0;
 			}
@@ -98,7 +98,7 @@ namespace Lucene.Net.Highlight
 		{
 			TokenStream ts = null;
 			
-			TermFreqVector tfv = (TermFreqVector) reader.GetTermFreqVector(docId, field);
+			ITermFreqVector tfv = (ITermFreqVector) reader.GetTermFreqVector(docId, field);
 			if (tfv != null)
 			{
 				if (tfv is TermPositionVector)
@@ -210,7 +210,7 @@ namespace Lucene.Net.Highlight
 		
 		public static TokenStream GetTokenStream(IndexReader reader, int docId, System.String field)
 		{
-			TermFreqVector tfv = (TermFreqVector) reader.GetTermFreqVector(docId, field);
+			ITermFreqVector tfv = (ITermFreqVector) reader.GetTermFreqVector(docId, field);
 			if (tfv == null)
 			{
 				throw new System.ArgumentException(field + " in doc #" + docId + "does not have any term position data stored");

Modified: incubator/lucene.net/trunk/src/contrib/Queries/FuzzyLikeThisQuery.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Queries/FuzzyLikeThisQuery.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Queries/FuzzyLikeThisQuery.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Queries/FuzzyLikeThisQuery.cs Wed Mar 21 06:04:26 2012
@@ -193,7 +193,7 @@ namespace Lucene.Net.Search
             TokenStream ts = analyzer.TokenStream(f.fieldName, new System.IO.StringReader(f.queryString));
             TermAttribute termAtt = ts.AddAttribute<TermAttribute>();
 
-            int corpusNumDocs = reader.NumDocs;
+            int corpusNumDocs = reader.GetNumDocs();
             Term internSavingTemplateTerm = new Term(f.fieldName); //optimization to avoid constructing new Term() objects
             HashSet<string> processedTerms = new HashSet<string>();
             while (ts.IncrementToken())

Modified: incubator/lucene.net/trunk/src/contrib/Queries/Similar/MoreLikeThis.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Queries/Similar/MoreLikeThis.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Queries/Similar/MoreLikeThis.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Queries/Similar/MoreLikeThis.cs Wed Mar 21 06:04:26 2012
@@ -19,12 +19,12 @@ using System;
 using System.Collections.Generic;
 using System.IO;
 using System.Linq;
+using Lucene.Net.Index;
 using Lucene.Net.Store;
 using Lucene.Net.Support;
 using Lucene.Net.Util;
 using IndexReader = Lucene.Net.Index.IndexReader;
 using Term = Lucene.Net.Index.Term;
-using TermFreqVector = Lucene.Net.Index.TermFreqVector;
 using BooleanClause = Lucene.Net.Search.BooleanClause;
 using DefaultSimilarity = Lucene.Net.Search.DefaultSimilarity;
 using TermQuery = Lucene.Net.Search.TermQuery;
@@ -403,7 +403,7 @@ namespace Lucene.Net.Search.Similar
         /// </param>
         public void SetMaxDocFreqPct(int maxPercentage)
         {
-            this.maxDocfreq = maxPercentage * ir.NumDocs / 100;
+            this.maxDocfreq = maxPercentage * ir.GetNumDocs() / 100;
         }
 
         /// <summary> Returns whether to boost terms in query based on "score" or not. The default is
@@ -675,7 +675,7 @@ namespace Lucene.Net.Search.Similar
         private PriorityQueue<object[]> CreateQueue(IDictionary<string,Int> words)
         {
             // have collected all words in doc and their freqs
-            int numDocs = ir.NumDocs;
+            int numDocs = ir.GetNumDocs();
             FreqQ res = new FreqQ(words.Count); // will order words by score
 
             var it = words.Keys.GetEnumerator();
@@ -777,7 +777,7 @@ namespace Lucene.Net.Search.Similar
             System.IO.StreamWriter o = temp_writer;
             FSDirectory dir = FSDirectory.Open(new DirectoryInfo(indexName));
             IndexReader r = IndexReader.Open(dir, true);
-            o.WriteLine("Open index " + indexName + " which has " + r.NumDocs + " docs");
+            o.WriteLine("Open index " + indexName + " which has " + r.GetNumDocs() + " docs");
 
             MoreLikeThis mlt = new MoreLikeThis(r);
 
@@ -830,7 +830,7 @@ namespace Lucene.Net.Search.Similar
             for (int i = 0; i < fieldNames.Length; i++)
             {
                 System.String fieldName = fieldNames[i];
-                TermFreqVector vector = ir.GetTermFreqVector(docNum, fieldName);
+                ITermFreqVector vector = ir.GetTermFreqVector(docNum, fieldName);
 
                 // field does not store term vector info
                 if (vector == null)
@@ -859,7 +859,7 @@ namespace Lucene.Net.Search.Similar
         /// </param>
         /// <param name="vector">List of terms and their frequencies for a doc/field
         /// </param>
-        private void AddTermFrequencies(IDictionary<string, Int> termFreqMap, TermFreqVector vector)
+        private void AddTermFrequencies(IDictionary<string, Int> termFreqMap, ITermFreqVector vector)
         {
             System.String[] terms = vector.GetTerms();
             int[] freqs = vector.GetTermFrequencies();

Modified: incubator/lucene.net/trunk/src/contrib/Similarity/Similar/MoreLikeThis.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Similarity/Similar/MoreLikeThis.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Similarity/Similar/MoreLikeThis.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Similarity/Similar/MoreLikeThis.cs Wed Mar 21 06:04:26 2012
@@ -18,12 +18,12 @@
 using System;
 using System.IO;
 using Lucene.Net.Analysis.Tokenattributes;
+using Lucene.Net.Index;
 using Lucene.Net.Search;
 using Lucene.Net.Store;
 using Lucene.Net.Util;
 using IndexReader = Lucene.Net.Index.IndexReader;
 using Term = Lucene.Net.Index.Term;
-using TermFreqVector = Lucene.Net.Index.TermFreqVector;
 using BooleanClause = Lucene.Net.Search.BooleanClause;
 using DefaultSimilarity = Lucene.Net.Search.DefaultSimilarity;
 using TermQuery = Lucene.Net.Search.TermQuery;
@@ -611,7 +611,7 @@ namespace Similarity.Net
         private PriorityQueue<object[]> CreateQueue(System.Collections.IDictionary words)
         {
             // have collected all words in doc and their freqs
-            int numDocs = ir.NumDocs;
+            int numDocs = ir.GetNumDocs();
             FreqQ res = new FreqQ(words.Count); // will order words by score
 			
             System.Collections.IEnumerator it = words.Keys.GetEnumerator();
@@ -706,7 +706,7 @@ namespace Similarity.Net
                               {AutoFlush = true};
             var dir = FSDirectory.Open(new DirectoryInfo(indexName));
             IndexReader r = IndexReader.Open(dir, true);
-            o.WriteLine("Open index " + indexName + " which has " + r.NumDocs + " docs");
+            o.WriteLine("Open index " + indexName + " which has " + r.GetNumDocs() + " docs");
 			
             MoreLikeThis mlt = new MoreLikeThis(r);
 			
@@ -760,7 +760,7 @@ namespace Similarity.Net
             for (int i = 0; i < fieldNames.Length; i++)
             {
                 System.String fieldName = fieldNames[i];
-                TermFreqVector vector = ir.GetTermFreqVector(docNum, fieldName);
+                ITermFreqVector vector = ir.GetTermFreqVector(docNum, fieldName);
 				
                 // field does not store term vector info
                 if (vector == null)
@@ -789,7 +789,7 @@ namespace Similarity.Net
         /// </param>
         /// <param name="vector">List of terms and their frequencies for a doc/field
         /// </param>
-        private void  AddTermFrequencies(System.Collections.IDictionary termFreqMap, TermFreqVector vector)
+        private void  AddTermFrequencies(System.Collections.IDictionary termFreqMap, ITermFreqVector vector)
         {
             System.String[] terms = vector.GetTerms();
             int[] freqs = vector.GetTermFrequencies();

Modified: incubator/lucene.net/trunk/src/contrib/Snowball/Lucene.Net/Analysis/Snowball/SnowballAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/Snowball/Lucene.Net/Analysis/Snowball/SnowballAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/Snowball/Lucene.Net/Analysis/Snowball/SnowballAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/Snowball/Lucene.Net/Analysis/Snowball/SnowballAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -102,7 +102,7 @@ namespace Lucene.Net.Analysis.Snowball
                 return TokenStream(fieldName, reader);
             }
 
-            SavedStreams streams = (SavedStreams)GetPreviousTokenStream();
+            SavedStreams streams = (SavedStreams)PreviousTokenStream;
             if (streams == null)
             {
                 streams = new SavedStreams();
@@ -113,7 +113,7 @@ namespace Lucene.Net.Analysis.Snowball
                     streams.result = new StopFilter(StopFilter.GetEnablePositionIncrementsVersionDefault(matchVersion),
                                                     streams.result, stopSet);
                 streams.result = new SnowballFilter(streams.result, name);
-                SetPreviousTokenStream(streams);
+                PreviousTokenStream = streams;
             }
             else
             {

Modified: incubator/lucene.net/trunk/src/contrib/SpellChecker/Spell/SpellChecker.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/contrib/SpellChecker/Spell/SpellChecker.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/contrib/SpellChecker/Spell/SpellChecker.cs (original)
+++ incubator/lucene.net/trunk/src/contrib/SpellChecker/Spell/SpellChecker.cs Wed Mar 21 06:04:26 2012
@@ -396,7 +396,7 @@ namespace SpellChecker.Net.Search.Spell
                 EnsureOpen();
                 Directory dir = this.spellindex;
                 IndexWriter writer = new IndexWriter(spellindex, new WhitespaceAnalyzer(), IndexWriter.MaxFieldLength.UNLIMITED);
-                writer.SetMergeFactor(mergeFactor);
+                writer.MergeFactor = mergeFactor;
                 writer.SetMaxBufferedDocs(ramMB);
 
                 System.Collections.IEnumerator iter = dict.GetWordsIterator();

Modified: incubator/lucene.net/trunk/src/core/Analysis/Analyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Analysis/Analyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Analysis/Analyzer.cs (original)
+++ incubator/lucene.net/trunk/src/core/Analysis/Analyzer.cs Wed Mar 21 06:04:26 2012
@@ -52,52 +52,31 @@ namespace Lucene.Net.Analysis
 		private CloseableThreadLocal<Object> tokenStreams = new CloseableThreadLocal<Object>();
 	    private bool isDisposed;
 
-        /// <summary>Used by Analyzers that implement reusableTokenStream
-        /// to save a TokenStream for later re-use by the same
-        /// thread. 
-        /// </summary>
+	    /// <summary>Used by Analyzers that implement reusableTokenStream
+	    /// to retrieve previously saved TokenStreams for re-use
+	    /// by the same thread. 
+	    /// </summary>
 	    protected internal virtual object PreviousTokenStream
 	    {
 	        get
-            {
-                if (tokenStreams == null)
-                {
-                    throw new AlreadyClosedException("this Analyzer is closed");
-                }
-                return tokenStreams.Get();
+	        {
+	            if (tokenStreams == null)
+	            {
+	                throw new AlreadyClosedException("this Analyzer is closed");
+	            }
+	            return tokenStreams.Get();
+	        }
+	        set
+	        {
+	            if (tokenStreams == null)
+	            {
+	                throw new AlreadyClosedException("this Analyzer is closed");
+	            }
+	            tokenStreams.Set(value);
 	        }
-
-            set
-            {
-                if (tokenStreams == null)
-                {
-                    throw new AlreadyClosedException("this Analyzer is closed");
-                }
-                tokenStreams.Set(value);
-            }
 	    }
 
-		/// <summary>Used by Analyzers that implement reusableTokenStream
-		/// to retrieve previously saved TokenStreams for re-use
-		/// by the same thread. 
-        /// </summary>
-        [Obsolete("Use PreviousTokenStream property instead")]
-		protected internal virtual System.Object GetPreviousTokenStream()
-		{
-		    return PreviousTokenStream;
-		}
-		
-		/// <summary>Used by Analyzers that implement reusableTokenStream
-		/// to save a TokenStream for later re-use by the same
-		/// thread. 
-		/// </summary>
-		[Obsolete("Use PreviousTokenStream property instead")]
-		protected internal virtual void  SetPreviousTokenStream(System.Object obj)
-		{
-		    PreviousTokenStream = obj;
-		}
-		
-        [Obsolete()]
+	    [Obsolete()]
 		protected internal bool overridesTokenStreamMethod = false;
 		
 		/// <deprecated> This is only present to preserve

Modified: incubator/lucene.net/trunk/src/core/Analysis/KeywordAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Analysis/KeywordAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Analysis/KeywordAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/src/core/Analysis/KeywordAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -42,11 +42,11 @@ namespace Lucene.Net.Analysis
 				// tokenStream but not reusableTokenStream
 				return TokenStream(fieldName, reader);
 			}
-			Tokenizer tokenizer = (Tokenizer) GetPreviousTokenStream();
+			Tokenizer tokenizer = (Tokenizer) PreviousTokenStream;
 			if (tokenizer == null)
 			{
 				tokenizer = new KeywordTokenizer(reader);
-				SetPreviousTokenStream(tokenizer);
+				PreviousTokenStream = tokenizer;
 			}
 			else
 				tokenizer.Reset(reader);

Modified: incubator/lucene.net/trunk/src/core/Analysis/NumericTokenStream.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Analysis/NumericTokenStream.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Analysis/NumericTokenStream.cs (original)
+++ incubator/lucene.net/trunk/src/core/Analysis/NumericTokenStream.cs Wed Mar 21 06:04:26 2012
@@ -246,7 +246,7 @@ namespace Lucene.Net.Analysis
 				
 			}
 			
-			typeAtt.SetType((shift == 0)?TOKEN_TYPE_FULL_PREC:TOKEN_TYPE_LOWER_PREC);
+			typeAtt.Type = (shift == 0)?TOKEN_TYPE_FULL_PREC:TOKEN_TYPE_LOWER_PREC;
 			posIncrAtt.PositionIncrement = (shift == 0)?1:0;
 			shift += precisionStep;
 			return true;

Modified: incubator/lucene.net/trunk/src/core/Analysis/PerFieldAnalyzerWrapper.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Analysis/PerFieldAnalyzerWrapper.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Analysis/PerFieldAnalyzerWrapper.cs (original)
+++ incubator/lucene.net/trunk/src/core/Analysis/PerFieldAnalyzerWrapper.cs Wed Mar 21 06:04:26 2012
@@ -122,10 +122,8 @@ namespace Lucene.Net.Analysis
 		/// <summary>Return the positionIncrementGap from the analyzer assigned to fieldName </summary>
 		public override int GetPositionIncrementGap(string fieldName)
 		{
-			Analyzer analyzer = analyzerMap[fieldName];
-			if (analyzer == null)
-				analyzer = defaultAnalyzer;
-			return analyzer.GetPositionIncrementGap(fieldName);
+			Analyzer analyzer = analyzerMap[fieldName] ?? defaultAnalyzer;
+		    return analyzer.GetPositionIncrementGap(fieldName);
 		}
 
         /// <summary> Return the offsetGap from the analyzer assigned to field </summary>

Modified: incubator/lucene.net/trunk/src/core/Analysis/PorterStemFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Analysis/PorterStemFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Analysis/PorterStemFilter.cs (original)
+++ incubator/lucene.net/trunk/src/core/Analysis/PorterStemFilter.cs Wed Mar 21 06:04:26 2012
@@ -57,7 +57,7 @@ namespace Lucene.Net.Analysis
 				return false;
 			
 			if (stemmer.Stem(termAtt.TermBuffer(), 0, termAtt.TermLength()))
-				termAtt.SetTermBuffer(stemmer.GetResultBuffer(), 0, stemmer.GetResultLength());
+				termAtt.SetTermBuffer(stemmer.ResultBuffer, 0, stemmer.ResultLength);
 			return true;
 		}
 	}



Mime
View raw message