C# Class Lucene.Net.Analysis.Compound.CompoundWordTokenFilterBase

Base class for decomposition token filters.

You must specify the required LuceneVersion compatibility when creating CompoundWordTokenFilterBase:

  • As of 3.1, CompoundWordTokenFilterBase correctly handles Unicode 4.0 supplementary characters in strings and char arrays provided as compound word dictionaries.
  • As of 4.4, CompoundWordTokenFilterBase doesn't update offsets.

Inheritance: TokenFilter
Afficher le fichier Open project: apache/lucenenet Class Usage Examples

Protected Properties

Свойство Type Description
dictionary CharArraySet
matchVersion LuceneVersion
maxSubwordSize int
minSubwordSize int
minWordSize int
offsetAtt IOffsetAttribute
onlyLongestMatch bool
termAtt ICharTermAttribute
tokens LinkedList

Méthodes publiques

Méthode Description
IncrementToken ( ) : bool
Reset ( ) : void

Méthodes protégées

Méthode Description
CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary ) : Lucene.Net.Analysis.Tokenattributes
CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary, bool onlyLongestMatch ) : Lucene.Net.Analysis.Tokenattributes
CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, bool onlyLongestMatch ) : Lucene.Net.Analysis.Tokenattributes
Decompose ( ) : void

Decomposes the current #termAtt and places CompoundToken instances in the #tokens list. The original token may not be placed in the list, as it is automatically passed through this filter.

Method Details

CompoundWordTokenFilterBase() protected méthode

protected CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary ) : Lucene.Net.Analysis.Tokenattributes
matchVersion LuceneVersion
input TokenStream
dictionary CharArraySet
Résultat Lucene.Net.Analysis.Tokenattributes

CompoundWordTokenFilterBase() protected méthode

protected CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary, bool onlyLongestMatch ) : Lucene.Net.Analysis.Tokenattributes
matchVersion LuceneVersion
input TokenStream
dictionary CharArraySet
onlyLongestMatch bool
Résultat Lucene.Net.Analysis.Tokenattributes

CompoundWordTokenFilterBase() protected méthode

protected CompoundWordTokenFilterBase ( LuceneVersion matchVersion, TokenStream input, CharArraySet dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, bool onlyLongestMatch ) : Lucene.Net.Analysis.Tokenattributes
matchVersion LuceneVersion
input TokenStream
dictionary CharArraySet
minWordSize int
minSubwordSize int
maxSubwordSize int
onlyLongestMatch bool
Résultat Lucene.Net.Analysis.Tokenattributes

Decompose() protected abstract méthode

Decomposes the current #termAtt and places CompoundToken instances in the #tokens list. The original token may not be placed in the list, as it is automatically passed through this filter.
protected abstract Decompose ( ) : void
Résultat void

IncrementToken() public final méthode

public final IncrementToken ( ) : bool
Résultat bool

Reset() public méthode

public Reset ( ) : void
Résultat void

Property Details

dictionary protected_oe property

protected CharArraySet dictionary
Résultat CharArraySet

matchVersion protected_oe property

protected LuceneVersion matchVersion
Résultat LuceneVersion

maxSubwordSize protected_oe property

protected int maxSubwordSize
Résultat int

minSubwordSize protected_oe property

protected int minSubwordSize
Résultat int

minWordSize protected_oe property

protected int minWordSize
Résultat int

offsetAtt protected_oe property

protected IOffsetAttribute offsetAtt
Résultat IOffsetAttribute

onlyLongestMatch protected_oe property

protected bool onlyLongestMatch
Résultat bool

termAtt protected_oe property

protected ICharTermAttribute termAtt
Résultat ICharTermAttribute

tokens protected_oe property

protected LinkedList tokens
Résultat LinkedList