Uses of Class
org.apache.lucene.util.AttributeSource

Packages that use AttributeSource
org.apache.lucene.analysis API and code to convert text into indexable/searchable tokens. 
org.apache.lucene.analysis.ar Analyzer for Arabic. 
org.apache.lucene.analysis.bg Analyzer for Bulgarian. 
org.apache.lucene.analysis.br Analyzer for Brazilian Portuguese. 
org.apache.lucene.analysis.cjk Analyzer for Chinese, Japanese, and Korean, which indexes bigrams (overlapping groups of two adjacent Han characters). 
org.apache.lucene.analysis.cn Analyzer for Chinese, which indexes unigrams (individual chinese characters). 
org.apache.lucene.analysis.cn.smart
Analyzer for Simplified Chinese, which indexes words. 
org.apache.lucene.analysis.compound A filter that decomposes compound words you find in many Germanic languages into the word parts. 
org.apache.lucene.analysis.cz Analyzer for Czech. 
org.apache.lucene.analysis.de Analyzer for German. 
org.apache.lucene.analysis.el Analyzer for Greek. 
org.apache.lucene.analysis.en Analyzer for English. 
org.apache.lucene.analysis.es Analyzer for Spanish. 
org.apache.lucene.analysis.fa Analyzer for Persian. 
org.apache.lucene.analysis.fi Analyzer for Finnish. 
org.apache.lucene.analysis.fr Analyzer for French. 
org.apache.lucene.analysis.gl Analyzer for Galician. 
org.apache.lucene.analysis.hi Analyzer for Hindi. 
org.apache.lucene.analysis.hu Analyzer for Hungarian. 
org.apache.lucene.analysis.icu Analysis components based on ICU 
org.apache.lucene.analysis.icu.segmentation Tokenizer that breaks text into words with the Unicode Text Segmentation algorithm. 
org.apache.lucene.analysis.id Analyzer for Indonesian. 
org.apache.lucene.analysis.in Analysis components for Indian languages. 
org.apache.lucene.analysis.it Analyzer for Italian. 
org.apache.lucene.analysis.miscellaneous Miscellaneous TokenStreams 
org.apache.lucene.analysis.ngram Character n-gram tokenizers and filters. 
org.apache.lucene.analysis.nl Analyzer for Dutch. 
org.apache.lucene.analysis.path   
org.apache.lucene.analysis.payloads
Provides various convenience classes for creating payloads on Tokens. 
org.apache.lucene.analysis.position Filter for assigning position increments. 
org.apache.lucene.analysis.pt Analyzer for Portuguese. 
org.apache.lucene.analysis.reverse Filter to reverse token text. 
org.apache.lucene.analysis.ru Analyzer for Russian. 
org.apache.lucene.analysis.shingle Word n-gram filters 
org.apache.lucene.analysis.sinks
Implementations of the SinkTokenizer that might be useful. 
org.apache.lucene.analysis.snowball TokenFilter and Analyzer implementations that use Snowball stemmers. 
org.apache.lucene.analysis.standard The org.apache.lucene.analysis.standard package contains three fast grammar-based tokenizers constructed with JFlex: 
org.apache.lucene.analysis.stempel Stempel: Algorithmic Stemmer 
org.apache.lucene.analysis.sv Analyzer for Swedish. 
org.apache.lucene.analysis.th Analyzer for Thai. 
org.apache.lucene.analysis.tr Analyzer for Turkish. 
org.apache.lucene.analysis.wikipedia Tokenizer that is aware of Wikipedia syntax. 
org.apache.lucene.collation CollationKeyFilter converts each token into its binary CollationKey using the provided Collator, and then encode the CollationKey as a String using IndexableBinaryStringTools, to allow it to be stored as an index term. 
org.apache.lucene.index Code to maintain and access indices. 
org.apache.lucene.queryParser.core.config Contains the base classes used to configure the query processing 
org.apache.lucene.queryParser.standard.config Standard Lucene Query Configuration 
org.apache.lucene.search.highlight The highlight package contains classes to provide "keyword in context" features typically used to highlight search terms in the text of results pages. 
org.apache.lucene.util Some utility classes. 
org.apache.lucene.wordnet This package uses synonyms defined by WordNet
 

Uses of AttributeSource in org.apache.lucene.analysis
 

Subclasses of AttributeSource in org.apache.lucene.analysis
 class ASCIIFoldingFilter
          This class converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if one exists.
 class CachingTokenFilter
          This class can be used if the token attributes of a TokenStream are intended to be consumed more than once.
 class CharTokenizer
          An abstract base class for simple, character-oriented tokenizers.
 class FilteringTokenFilter
          Abstract base class for TokenFilters that may remove tokens.
 class ISOLatin1AccentFilter
          Deprecated. If you build a new index, use ASCIIFoldingFilter which covers a superset of Latin 1. This class is included for use with existing indexes and will be removed in a future release (possibly Lucene 4.0).
 class KeywordMarkerFilter
          Marks terms as keywords via the KeywordAttribute.
 class KeywordTokenizer
          Emits the entire input as a single token.
 class LengthFilter
          Removes words that are too long or too short from the stream.
 class LetterTokenizer
          A LetterTokenizer is a tokenizer that divides text at non-letters.
 class LimitTokenCountFilter
          This TokenFilter limits the number of tokens while indexing.
 class LowerCaseFilter
          Normalizes token text to lower case.
 class LowerCaseTokenizer
          LowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together.
 class NumericTokenStream
          Expert: This class provides a TokenStream for indexing numeric values that can be used by NumericRangeQuery or NumericRangeFilter.
 class PorterStemFilter
          Transforms the token stream as per the Porter stemming algorithm.
 class StopFilter
          Removes stop words from a token stream.
 class TeeSinkTokenFilter
          This TokenFilter provides the ability to set aside attribute states that have already been analyzed.
static class TeeSinkTokenFilter.SinkTokenStream
           
 class TokenFilter
          A TokenFilter is a TokenStream whose input is another TokenStream.
 class Tokenizer
          A Tokenizer is a TokenStream whose input is a Reader.
 class TokenStream
          A TokenStream enumerates the sequence of tokens, either from Fields of a Document or from query text.
 class WhitespaceTokenizer
          A WhitespaceTokenizer is a tokenizer that divides text at whitespace.
 

Methods in org.apache.lucene.analysis with parameters of type AttributeSource
abstract  boolean TeeSinkTokenFilter.SinkFilter.accept(AttributeSource source)
          Returns true, iff the current state of the passed-in AttributeSource shall be stored in the sink.
 

Constructors in org.apache.lucene.analysis with parameters of type AttributeSource
CharTokenizer(AttributeSource source, Reader input)
          Deprecated. use CharTokenizer.CharTokenizer(Version, AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
CharTokenizer(Version matchVersion, AttributeSource source, Reader input)
          Creates a new CharTokenizer instance
KeywordTokenizer(AttributeSource source, Reader input, int bufferSize)
           
LetterTokenizer(AttributeSource source, Reader in)
          Deprecated. use LetterTokenizer.LetterTokenizer(Version, AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
LetterTokenizer(Version matchVersion, AttributeSource source, Reader in)
          Construct a new LetterTokenizer using a given AttributeSource.
LowerCaseTokenizer(AttributeSource source, Reader in)
          Deprecated. use LowerCaseTokenizer.LowerCaseTokenizer(AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
LowerCaseTokenizer(Version matchVersion, AttributeSource source, Reader in)
          Construct a new LowerCaseTokenizer using a given AttributeSource.
NumericTokenStream(AttributeSource source, int precisionStep)
          Expert: Creates a token stream for numeric values with the specified precisionStep using the given AttributeSource.
Tokenizer(AttributeSource source)
          Construct a token stream processing the given input using the given AttributeSource.
Tokenizer(AttributeSource source, Reader input)
          Construct a token stream processing the given input using the given AttributeSource.
TokenStream(AttributeSource input)
          A TokenStream that uses the same attributes as the supplied one.
WhitespaceTokenizer(AttributeSource source, Reader in)
          Deprecated. use WhitespaceTokenizer.WhitespaceTokenizer(Version, AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
WhitespaceTokenizer(Version matchVersion, AttributeSource source, Reader in)
          Construct a new WhitespaceTokenizer using a given AttributeSource.
 

Uses of AttributeSource in org.apache.lucene.analysis.ar
 

Subclasses of AttributeSource in org.apache.lucene.analysis.ar
 class ArabicLetterTokenizer
          Deprecated. (3.1) Use StandardTokenizer instead.
 class ArabicNormalizationFilter
          A TokenFilter that applies ArabicNormalizer to normalize the orthography.
 class ArabicStemFilter
          A TokenFilter that applies ArabicStemmer to stem Arabic words..
 

Constructors in org.apache.lucene.analysis.ar with parameters of type AttributeSource
ArabicLetterTokenizer(AttributeSource source, Reader in)
          Deprecated. use ArabicLetterTokenizer.ArabicLetterTokenizer(Version, AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
ArabicLetterTokenizer(Version matchVersion, AttributeSource source, Reader in)
          Deprecated. Construct a new ArabicLetterTokenizer using a given AttributeSource.
 

Uses of AttributeSource in org.apache.lucene.analysis.bg
 

Subclasses of AttributeSource in org.apache.lucene.analysis.bg
 class BulgarianStemFilter
          A TokenFilter that applies BulgarianStemmer to stem Bulgarian words.
 

Uses of AttributeSource in org.apache.lucene.analysis.br
 

Subclasses of AttributeSource in org.apache.lucene.analysis.br
 class BrazilianStemFilter
          A TokenFilter that applies BrazilianStemmer.
 

Uses of AttributeSource in org.apache.lucene.analysis.cjk
 

Subclasses of AttributeSource in org.apache.lucene.analysis.cjk
 class CJKTokenizer
          CJKTokenizer is designed for Chinese, Japanese, and Korean languages.
 

Constructors in org.apache.lucene.analysis.cjk with parameters of type AttributeSource
CJKTokenizer(AttributeSource source, Reader in)
           
 

Uses of AttributeSource in org.apache.lucene.analysis.cn
 

Subclasses of AttributeSource in org.apache.lucene.analysis.cn
 class ChineseFilter
          Deprecated. Use StopFilter instead, which has the same functionality. This filter will be removed in Lucene 5.0
 class ChineseTokenizer
          Deprecated. Use StandardTokenizer instead, which has the same functionality. This filter will be removed in Lucene 5.0
 

Constructors in org.apache.lucene.analysis.cn with parameters of type AttributeSource
ChineseTokenizer(AttributeSource source, Reader in)
          Deprecated.  
 

Uses of AttributeSource in org.apache.lucene.analysis.cn.smart
 

Subclasses of AttributeSource in org.apache.lucene.analysis.cn.smart
 class SentenceTokenizer
          Tokenizes input text into sentences.
 class WordTokenFilter
          A TokenFilter that breaks sentences into words.
 

Constructors in org.apache.lucene.analysis.cn.smart with parameters of type AttributeSource
SentenceTokenizer(AttributeSource source, Reader reader)
           
 

Uses of AttributeSource in org.apache.lucene.analysis.compound
 

Subclasses of AttributeSource in org.apache.lucene.analysis.compound
 class CompoundWordTokenFilterBase
          Base class for decomposition token filters.
 class DictionaryCompoundWordTokenFilter
          A TokenFilter that decomposes compound words found in many Germanic languages.
 class HyphenationCompoundWordTokenFilter
          A TokenFilter that decomposes compound words found in many Germanic languages.
 

Uses of AttributeSource in org.apache.lucene.analysis.cz
 

Subclasses of AttributeSource in org.apache.lucene.analysis.cz
 class CzechStemFilter
          A TokenFilter that applies CzechStemmer to stem Czech words.
 

Uses of AttributeSource in org.apache.lucene.analysis.de
 

Subclasses of AttributeSource in org.apache.lucene.analysis.de
 class GermanLightStemFilter
          A TokenFilter that applies GermanLightStemmer to stem German words.
 class GermanMinimalStemFilter
          A TokenFilter that applies GermanMinimalStemmer to stem German words.
 class GermanStemFilter
          A TokenFilter that stems German words.
 

Uses of AttributeSource in org.apache.lucene.analysis.el
 

Subclasses of AttributeSource in org.apache.lucene.analysis.el
 class GreekLowerCaseFilter
          Normalizes token text to lower case, removes some Greek diacritics, and standardizes final sigma to sigma.
 class GreekStemFilter
          A TokenFilter that applies GreekStemmer to stem Greek words.
 

Uses of AttributeSource in org.apache.lucene.analysis.en
 

Subclasses of AttributeSource in org.apache.lucene.analysis.en
 class EnglishMinimalStemFilter
          A TokenFilter that applies EnglishMinimalStemmer to stem English words.
 class EnglishPossessiveFilter
          TokenFilter that removes possessives (trailing 's) from words.
 

Uses of AttributeSource in org.apache.lucene.analysis.es
 

Subclasses of AttributeSource in org.apache.lucene.analysis.es
 class SpanishLightStemFilter
          A TokenFilter that applies SpanishLightStemmer to stem Spanish words.
 

Uses of AttributeSource in org.apache.lucene.analysis.fa
 

Subclasses of AttributeSource in org.apache.lucene.analysis.fa
 class PersianNormalizationFilter
          A TokenFilter that applies PersianNormalizer to normalize the orthography.
 

Uses of AttributeSource in org.apache.lucene.analysis.fi
 

Subclasses of AttributeSource in org.apache.lucene.analysis.fi
 class FinnishLightStemFilter
          A TokenFilter that applies FinnishLightStemmer to stem Finnish words.
 

Uses of AttributeSource in org.apache.lucene.analysis.fr
 

Subclasses of AttributeSource in org.apache.lucene.analysis.fr
 class ElisionFilter
          Removes elisions from a TokenStream.
 class FrenchLightStemFilter
          A TokenFilter that applies FrenchLightStemmer to stem French words.
 class FrenchMinimalStemFilter
          A TokenFilter that applies FrenchMinimalStemmer to stem French words.
 class FrenchStemFilter
          Deprecated. Use SnowballFilter with FrenchStemmer instead, which has the same functionality. This filter will be removed in Lucene 5.0
 

Uses of AttributeSource in org.apache.lucene.analysis.gl
 

Subclasses of AttributeSource in org.apache.lucene.analysis.gl
 class GalicianStemFilter
          A TokenFilter that applies GalicianStemmer to stem Galician words.
 

Uses of AttributeSource in org.apache.lucene.analysis.hi
 

Subclasses of AttributeSource in org.apache.lucene.analysis.hi
 class HindiNormalizationFilter
          A TokenFilter that applies HindiNormalizer to normalize the orthography.
 class HindiStemFilter
          A TokenFilter that applies HindiStemmer to stem Hindi words.
 

Uses of AttributeSource in org.apache.lucene.analysis.hu
 

Subclasses of AttributeSource in org.apache.lucene.analysis.hu
 class HungarianLightStemFilter
          A TokenFilter that applies HungarianLightStemmer to stem Hungarian words.
 

Uses of AttributeSource in org.apache.lucene.analysis.icu
 

Subclasses of AttributeSource in org.apache.lucene.analysis.icu
 class ICUFoldingFilter
          A TokenFilter that applies search term folding to Unicode text, applying foldings from UTR#30 Character Foldings.
 class ICUNormalizer2Filter
          Normalize token text with ICU's Normalizer2
 class ICUTransformFilter
          A TokenFilter that transforms text with ICU.
 

Uses of AttributeSource in org.apache.lucene.analysis.icu.segmentation
 

Subclasses of AttributeSource in org.apache.lucene.analysis.icu.segmentation
 class ICUTokenizer
          Breaks text into words according to UAX #29: Unicode Text Segmentation (http://www.unicode.org/reports/tr29/)
 

Uses of AttributeSource in org.apache.lucene.analysis.id
 

Subclasses of AttributeSource in org.apache.lucene.analysis.id
 class IndonesianStemFilter
          A TokenFilter that applies IndonesianStemmer to stem Indonesian words.
 

Uses of AttributeSource in org.apache.lucene.analysis.in
 

Subclasses of AttributeSource in org.apache.lucene.analysis.in
 class IndicNormalizationFilter
          A TokenFilter that applies IndicNormalizer to normalize text in Indian Languages.
 class IndicTokenizer
          Simple Tokenizer for text in Indian Languages.
 

Constructors in org.apache.lucene.analysis.in with parameters of type AttributeSource
IndicTokenizer(Version matchVersion, AttributeSource source, Reader input)
           
 

Uses of AttributeSource in org.apache.lucene.analysis.it
 

Subclasses of AttributeSource in org.apache.lucene.analysis.it
 class ItalianLightStemFilter
          A TokenFilter that applies ItalianLightStemmer to stem Italian words.
 

Uses of AttributeSource in org.apache.lucene.analysis.miscellaneous
 

Subclasses of AttributeSource in org.apache.lucene.analysis.miscellaneous
 class EmptyTokenStream
          An always exhausted token stream.
 class PrefixAndSuffixAwareTokenFilter
          Links two PrefixAwareTokenFilter.
 class PrefixAwareTokenFilter
          Joins two token streams and leaves the last token of the first stream available to be used when updating the token values in the second stream based on that token.
 class SingleTokenTokenStream
          A TokenStream containing a single token.
 class StemmerOverrideFilter
          Provides the ability to override any KeywordAttribute aware stemmer with custom dictionary-based stemming.
 

Uses of AttributeSource in org.apache.lucene.analysis.ngram
 

Subclasses of AttributeSource in org.apache.lucene.analysis.ngram
 class EdgeNGramTokenFilter
          Tokenizes the given token into n-grams of given size(s).
 class EdgeNGramTokenizer
          Tokenizes the input from an edge into n-grams of given size(s).
 class NGramTokenFilter
          Tokenizes the input into n-grams of the given size(s).
 class NGramTokenizer
          Tokenizes the input into n-grams of the given size(s).
 

Constructors in org.apache.lucene.analysis.ngram with parameters of type AttributeSource
EdgeNGramTokenizer(AttributeSource source, Reader input, EdgeNGramTokenizer.Side side, int minGram, int maxGram)
          Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range
EdgeNGramTokenizer(AttributeSource source, Reader input, String sideLabel, int minGram, int maxGram)
          Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range
NGramTokenizer(AttributeSource source, Reader input, int minGram, int maxGram)
          Creates NGramTokenizer with given min and max n-grams.
 

Uses of AttributeSource in org.apache.lucene.analysis.nl
 

Subclasses of AttributeSource in org.apache.lucene.analysis.nl
 class DutchStemFilter
          Deprecated. Use SnowballFilter with DutchStemmer instead, which has the same functionality. This filter will be removed in Lucene 5.0
 

Uses of AttributeSource in org.apache.lucene.analysis.path
 

Subclasses of AttributeSource in org.apache.lucene.analysis.path
 class PathHierarchyTokenizer
          Take something like:
 

Uses of AttributeSource in org.apache.lucene.analysis.payloads
 

Subclasses of AttributeSource in org.apache.lucene.analysis.payloads
 class DelimitedPayloadTokenFilter
          Characters before the delimiter are the "token", those after are the payload.
 class NumericPayloadTokenFilter
          Assigns a payload to a token based on the Token.type()
 class TokenOffsetPayloadTokenFilter
          Adds the Token.setStartOffset(int) and Token.setEndOffset(int) First 4 bytes are the start
 class TypeAsPayloadTokenFilter
          Makes the Token.type() a payload.
 

Uses of AttributeSource in org.apache.lucene.analysis.position
 

Subclasses of AttributeSource in org.apache.lucene.analysis.position
 class PositionFilter
          Set the positionIncrement of all tokens to the "positionIncrement", except the first return token which retains its original positionIncrement value.
 

Uses of AttributeSource in org.apache.lucene.analysis.pt
 

Subclasses of AttributeSource in org.apache.lucene.analysis.pt
 class PortugueseLightStemFilter
          A TokenFilter that applies PortugueseLightStemmer to stem Portuguese words.
 class PortugueseMinimalStemFilter
          A TokenFilter that applies PortugueseMinimalStemmer to stem Portuguese words.
 class PortugueseStemFilter
          A TokenFilter that applies PortugueseStemmer to stem Portuguese words.
 

Uses of AttributeSource in org.apache.lucene.analysis.reverse
 

Subclasses of AttributeSource in org.apache.lucene.analysis.reverse
 class ReverseStringFilter
          Reverse token string, for example "country" => "yrtnuoc".
 

Uses of AttributeSource in org.apache.lucene.analysis.ru
 

Subclasses of AttributeSource in org.apache.lucene.analysis.ru
 class RussianLetterTokenizer
          Deprecated. Use StandardTokenizer instead, which has the same functionality. This filter will be removed in Lucene 5.0
 class RussianLightStemFilter
          A TokenFilter that applies RussianLightStemmer to stem Russian words.
 class RussianLowerCaseFilter
          Deprecated. Use LowerCaseFilter instead, which has the same functionality. This filter will be removed in Lucene 4.0
 class RussianStemFilter
          Deprecated. Use SnowballFilter with RussianStemmer instead, which has the same functionality. This filter will be removed in Lucene 4.0
 

Constructors in org.apache.lucene.analysis.ru with parameters of type AttributeSource
RussianLetterTokenizer(AttributeSource source, Reader in)
          Deprecated. use RussianLetterTokenizer.RussianLetterTokenizer(Version, AttributeSource, Reader) instead. This will be removed in Lucene 4.0.
RussianLetterTokenizer(Version matchVersion, AttributeSource source, Reader in)
          Deprecated. Construct a new RussianLetterTokenizer using a given AttributeSource.
 

Uses of AttributeSource in org.apache.lucene.analysis.shingle
 

Subclasses of AttributeSource in org.apache.lucene.analysis.shingle
 class ShingleFilter
          A ShingleFilter constructs shingles (token n-grams) from a token stream.
 class ShingleMatrixFilter
          Deprecated. Will be removed in Lucene 4.0. This filter is unmaintained and might not behave correctly if used with custom Attributes, i.e. Attributes other than the ones located in org.apache.lucene.analysis.tokenattributes. It also uses hardcoded payload encoders which makes it not easily adaptable to other use-cases.
 

Uses of AttributeSource in org.apache.lucene.analysis.sinks
 

Methods in org.apache.lucene.analysis.sinks with parameters of type AttributeSource
 boolean TokenTypeSinkFilter.accept(AttributeSource source)
           
 boolean TokenRangeSinkFilter.accept(AttributeSource source)
           
 boolean DateRecognizerSinkFilter.accept(AttributeSource source)
           
 

Uses of AttributeSource in org.apache.lucene.analysis.snowball
 

Subclasses of AttributeSource in org.apache.lucene.analysis.snowball
 class SnowballFilter
          A filter that stems words using a Snowball-generated stemmer.
 

Uses of AttributeSource in org.apache.lucene.analysis.standard
 

Subclasses of AttributeSource in org.apache.lucene.analysis.standard
 class ClassicFilter
          Normalizes tokens extracted with ClassicTokenizer.
 class ClassicTokenizer
          A grammar-based tokenizer constructed with JFlex
 class StandardFilter
          Normalizes tokens extracted with StandardTokenizer.
 class StandardTokenizer
          A grammar-based tokenizer constructed with JFlex.
 class UAX29URLEmailTokenizer
          This class implements Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29 URLs and email addresses are also tokenized according to the relevant RFCs.
 

Constructors in org.apache.lucene.analysis.standard with parameters of type AttributeSource
ClassicTokenizer(Version matchVersion, AttributeSource source, Reader input)
          Creates a new ClassicTokenizer with a given AttributeSource.
StandardTokenizer(Version matchVersion, AttributeSource source, Reader input)
          Creates a new StandardTokenizer with a given AttributeSource.
UAX29URLEmailTokenizer(AttributeSource source, Reader input)
           
 

Uses of AttributeSource in org.apache.lucene.analysis.stempel
 

Subclasses of AttributeSource in org.apache.lucene.analysis.stempel
 class StempelFilter
          Transforms the token stream as per the stemming algorithm.
 

Uses of AttributeSource in org.apache.lucene.analysis.sv
 

Subclasses of AttributeSource in org.apache.lucene.analysis.sv
 class SwedishLightStemFilter
          A TokenFilter that applies SwedishLightStemmer to stem Swedish words.
 

Uses of AttributeSource in org.apache.lucene.analysis.th
 

Subclasses of AttributeSource in org.apache.lucene.analysis.th
 class ThaiWordFilter
          TokenFilter that use BreakIterator to break each Token that is Thai into separate Token(s) for each Thai word.
 

Uses of AttributeSource in org.apache.lucene.analysis.tr
 

Subclasses of AttributeSource in org.apache.lucene.analysis.tr
 class TurkishLowerCaseFilter
          Normalizes Turkish token text to lower case.
 

Uses of AttributeSource in org.apache.lucene.analysis.wikipedia
 

Subclasses of AttributeSource in org.apache.lucene.analysis.wikipedia
 class WikipediaTokenizer
          Extension of StandardTokenizer that is aware of Wikipedia syntax.
 

Constructors in org.apache.lucene.analysis.wikipedia with parameters of type AttributeSource
WikipediaTokenizer(AttributeSource source, Reader input, int tokenOutput, Set<String> untokenizedTypes)
          Creates a new instance of the WikipediaTokenizer.
 

Uses of AttributeSource in org.apache.lucene.collation
 

Subclasses of AttributeSource in org.apache.lucene.collation
 class CollationKeyFilter
           Converts each token into its CollationKey, and then encodes the CollationKey with IndexableBinaryStringTools, to allow it to be stored as an index term.
 class ICUCollationKeyFilter
           Converts each token into its CollationKey, and then encodes the CollationKey with IndexableBinaryStringTools, to allow it to be stored as an index term.
 

Uses of AttributeSource in org.apache.lucene.index
 

Methods in org.apache.lucene.index that return AttributeSource
 AttributeSource FieldInvertState.getAttributeSource()
           
 

Uses of AttributeSource in org.apache.lucene.queryParser.core.config
 

Subclasses of AttributeSource in org.apache.lucene.queryParser.core.config
 class FieldConfig
          This class represents a field configuration.
 class QueryConfigHandler
          This class can be used to hold any query configuration and no field configuration.
 

Uses of AttributeSource in org.apache.lucene.queryParser.standard.config
 

Subclasses of AttributeSource in org.apache.lucene.queryParser.standard.config
 class StandardQueryConfigHandler
          This query configuration handler is used for almost every processor defined in the StandardQueryNodeProcessorPipeline processor pipeline.
 

Uses of AttributeSource in org.apache.lucene.search.highlight
 

Subclasses of AttributeSource in org.apache.lucene.search.highlight
 class TokenStreamFromTermPositionVector
           
 

Uses of AttributeSource in org.apache.lucene.util
 

Methods in org.apache.lucene.util that return AttributeSource
 AttributeSource AttributeSource.cloneAttributes()
          Performs a clone of all AttributeImpl instances returned in a new AttributeSource instance.
 

Methods in org.apache.lucene.util with parameters of type AttributeSource
 void AttributeSource.copyTo(AttributeSource target)
          Copies the contents of this AttributeSource to the given target AttributeSource.
 

Constructors in org.apache.lucene.util with parameters of type AttributeSource
AttributeSource(AttributeSource input)
          An AttributeSource that uses the same attributes as the supplied one.
 

Uses of AttributeSource in org.apache.lucene.wordnet
 

Subclasses of AttributeSource in org.apache.lucene.wordnet
 class SynonymTokenFilter
          Injects additional tokens for synonyms of token terms fetched from the underlying child stream; the child stream must deliver lowercase tokens for synonyms to be found.
 



Copyright © 2000-2011 Apache Software Foundation. All Rights Reserved.