|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use AttributeSource | |
---|---|
org.apache.lucene.analysis | API and code to convert text into indexable/searchable tokens. |
org.apache.lucene.analysis.ar | Analyzer for Arabic. |
org.apache.lucene.analysis.bg | Analyzer for Bulgarian. |
org.apache.lucene.analysis.br | Analyzer for Brazilian Portuguese. |
org.apache.lucene.analysis.cjk | Analyzer for Chinese, Japanese, and Korean, which indexes bigrams (overlapping groups of two adjacent Han characters). |
org.apache.lucene.analysis.cn | Analyzer for Chinese, which indexes unigrams (individual chinese characters). |
org.apache.lucene.analysis.cn.smart |
Analyzer for Simplified Chinese, which indexes words. |
org.apache.lucene.analysis.compound | A filter that decomposes compound words you find in many Germanic languages into the word parts. |
org.apache.lucene.analysis.cz | Analyzer for Czech. |
org.apache.lucene.analysis.de | Analyzer for German. |
org.apache.lucene.analysis.el | Analyzer for Greek. |
org.apache.lucene.analysis.en | Analyzer for English. |
org.apache.lucene.analysis.es | Analyzer for Spanish. |
org.apache.lucene.analysis.fa | Analyzer for Persian. |
org.apache.lucene.analysis.fi | Analyzer for Finnish. |
org.apache.lucene.analysis.fr | Analyzer for French. |
org.apache.lucene.analysis.gl | Analyzer for Galician. |
org.apache.lucene.analysis.hi | Analyzer for Hindi. |
org.apache.lucene.analysis.hu | Analyzer for Hungarian. |
org.apache.lucene.analysis.icu | Analysis components based on ICU |
org.apache.lucene.analysis.icu.segmentation | Tokenizer that breaks text into words with the Unicode Text Segmentation algorithm. |
org.apache.lucene.analysis.id | Analyzer for Indonesian. |
org.apache.lucene.analysis.in | Analysis components for Indian languages. |
org.apache.lucene.analysis.it | Analyzer for Italian. |
org.apache.lucene.analysis.miscellaneous | Miscellaneous TokenStreams |
org.apache.lucene.analysis.ngram | Character n-gram tokenizers and filters. |
org.apache.lucene.analysis.nl | Analyzer for Dutch. |
org.apache.lucene.analysis.path | |
org.apache.lucene.analysis.payloads | Provides various convenience classes for creating payloads on Tokens. |
org.apache.lucene.analysis.position | Filter for assigning position increments. |
org.apache.lucene.analysis.pt | Analyzer for Portuguese. |
org.apache.lucene.analysis.reverse | Filter to reverse token text. |
org.apache.lucene.analysis.ru | Analyzer for Russian. |
org.apache.lucene.analysis.shingle | Word n-gram filters |
org.apache.lucene.analysis.sinks | Implementations of the SinkTokenizer that might be useful. |
org.apache.lucene.analysis.snowball | TokenFilter and Analyzer implementations that use Snowball
stemmers. |
org.apache.lucene.analysis.standard | The org.apache.lucene.analysis.standard package contains three
fast grammar-based tokenizers constructed with JFlex: |
org.apache.lucene.analysis.stempel | Stempel: Algorithmic Stemmer |
org.apache.lucene.analysis.sv | Analyzer for Swedish. |
org.apache.lucene.analysis.th | Analyzer for Thai. |
org.apache.lucene.analysis.tr | Analyzer for Turkish. |
org.apache.lucene.analysis.wikipedia | Tokenizer that is aware of Wikipedia syntax. |
org.apache.lucene.collation |
CollationKeyFilter
converts each token into its binary CollationKey using the
provided Collator , and then encode the CollationKey
as a String using
IndexableBinaryStringTools , to allow it to be
stored as an index term. |
org.apache.lucene.index | Code to maintain and access indices. |
org.apache.lucene.queryParser.core.config | Contains the base classes used to configure the query processing |
org.apache.lucene.queryParser.standard.config | Standard Lucene Query Configuration |
org.apache.lucene.search.highlight | The highlight package contains classes to provide "keyword in context" features typically used to highlight search terms in the text of results pages. |
org.apache.lucene.util | Some utility classes. |
org.apache.lucene.wordnet | This package uses synonyms defined by WordNet. |
Uses of AttributeSource in org.apache.lucene.analysis |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis | |
---|---|
class |
ASCIIFoldingFilter
This class converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if one exists. |
class |
CachingTokenFilter
This class can be used if the token attributes of a TokenStream are intended to be consumed more than once. |
class |
CharTokenizer
An abstract base class for simple, character-oriented tokenizers. |
class |
FilteringTokenFilter
Abstract base class for TokenFilters that may remove tokens. |
class |
ISOLatin1AccentFilter
Deprecated. If you build a new index, use ASCIIFoldingFilter
which covers a superset of Latin 1.
This class is included for use with existing
indexes and will be removed in a future release (possibly Lucene 4.0). |
class |
KeywordMarkerFilter
Marks terms as keywords via the KeywordAttribute . |
class |
KeywordTokenizer
Emits the entire input as a single token. |
class |
LengthFilter
Removes words that are too long or too short from the stream. |
class |
LetterTokenizer
A LetterTokenizer is a tokenizer that divides text at non-letters. |
class |
LimitTokenCountFilter
This TokenFilter limits the number of tokens while indexing. |
class |
LowerCaseFilter
Normalizes token text to lower case. |
class |
LowerCaseTokenizer
LowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together. |
class |
NumericTokenStream
Expert: This class provides a TokenStream
for indexing numeric values that can be used by NumericRangeQuery or NumericRangeFilter . |
class |
PorterStemFilter
Transforms the token stream as per the Porter stemming algorithm. |
class |
StopFilter
Removes stop words from a token stream. |
class |
TeeSinkTokenFilter
This TokenFilter provides the ability to set aside attribute states that have already been analyzed. |
static class |
TeeSinkTokenFilter.SinkTokenStream
|
class |
TokenFilter
A TokenFilter is a TokenStream whose input is another TokenStream. |
class |
Tokenizer
A Tokenizer is a TokenStream whose input is a Reader. |
class |
TokenStream
A TokenStream enumerates the sequence of tokens, either from
Field s of a Document or from query text. |
class |
WhitespaceTokenizer
A WhitespaceTokenizer is a tokenizer that divides text at whitespace. |
Methods in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
abstract boolean |
TeeSinkTokenFilter.SinkFilter.accept(AttributeSource source)
Returns true, iff the current state of the passed-in AttributeSource shall be stored
in the sink. |
Constructors in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
CharTokenizer(AttributeSource source,
Reader input)
Deprecated. use CharTokenizer.CharTokenizer(Version, AttributeSource, Reader) instead. This will be
removed in Lucene 4.0. |
|
CharTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new CharTokenizer instance |
|
KeywordTokenizer(AttributeSource source,
Reader input,
int bufferSize)
|
|
LetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use LetterTokenizer.LetterTokenizer(Version, AttributeSource, Reader) instead.
This will be removed in Lucene 4.0. |
|
LetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new LetterTokenizer using a given AttributeSource . |
|
LowerCaseTokenizer(AttributeSource source,
Reader in)
Deprecated. use LowerCaseTokenizer.LowerCaseTokenizer(AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
LowerCaseTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new LowerCaseTokenizer using a given AttributeSource . |
|
NumericTokenStream(AttributeSource source,
int precisionStep)
Expert: Creates a token stream for numeric values with the specified precisionStep using the given AttributeSource . |
|
Tokenizer(AttributeSource source)
Construct a token stream processing the given input using the given AttributeSource. |
|
Tokenizer(AttributeSource source,
Reader input)
Construct a token stream processing the given input using the given AttributeSource. |
|
TokenStream(AttributeSource input)
A TokenStream that uses the same attributes as the supplied one. |
|
WhitespaceTokenizer(AttributeSource source,
Reader in)
Deprecated. use WhitespaceTokenizer.WhitespaceTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
WhitespaceTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new WhitespaceTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.ar |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ar | |
---|---|
class |
ArabicLetterTokenizer
Deprecated. (3.1) Use StandardTokenizer instead. |
class |
ArabicNormalizationFilter
A TokenFilter that applies ArabicNormalizer to normalize the orthography. |
class |
ArabicStemFilter
A TokenFilter that applies ArabicStemmer to stem Arabic words.. |
Constructors in org.apache.lucene.analysis.ar with parameters of type AttributeSource | |
---|---|
ArabicLetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use ArabicLetterTokenizer.ArabicLetterTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
ArabicLetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Deprecated. Construct a new ArabicLetterTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.bg |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.bg | |
---|---|
class |
BulgarianStemFilter
A TokenFilter that applies BulgarianStemmer to stem Bulgarian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.br |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.br | |
---|---|
class |
BrazilianStemFilter
A TokenFilter that applies BrazilianStemmer . |
Uses of AttributeSource in org.apache.lucene.analysis.cjk |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cjk | |
---|---|
class |
CJKTokenizer
CJKTokenizer is designed for Chinese, Japanese, and Korean languages. |
Constructors in org.apache.lucene.analysis.cjk with parameters of type AttributeSource | |
---|---|
CJKTokenizer(AttributeSource source,
Reader in)
|
Uses of AttributeSource in org.apache.lucene.analysis.cn |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn | |
---|---|
class |
ChineseFilter
Deprecated. Use StopFilter instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
class |
ChineseTokenizer
Deprecated. Use StandardTokenizer instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
Constructors in org.apache.lucene.analysis.cn with parameters of type AttributeSource | |
---|---|
ChineseTokenizer(AttributeSource source,
Reader in)
Deprecated. |
Uses of AttributeSource in org.apache.lucene.analysis.cn.smart |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn.smart | |
---|---|
class |
SentenceTokenizer
Tokenizes input text into sentences. |
class |
WordTokenFilter
A TokenFilter that breaks sentences into words. |
Constructors in org.apache.lucene.analysis.cn.smart with parameters of type AttributeSource | |
---|---|
SentenceTokenizer(AttributeSource source,
Reader reader)
|
Uses of AttributeSource in org.apache.lucene.analysis.compound |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.compound | |
---|---|
class |
CompoundWordTokenFilterBase
Base class for decomposition token filters. |
class |
DictionaryCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
class |
HyphenationCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
Uses of AttributeSource in org.apache.lucene.analysis.cz |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cz | |
---|---|
class |
CzechStemFilter
A TokenFilter that applies CzechStemmer to stem Czech words. |
Uses of AttributeSource in org.apache.lucene.analysis.de |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.de | |
---|---|
class |
GermanLightStemFilter
A TokenFilter that applies GermanLightStemmer to stem German
words. |
class |
GermanMinimalStemFilter
A TokenFilter that applies GermanMinimalStemmer to stem German
words. |
class |
GermanStemFilter
A TokenFilter that stems German words. |
Uses of AttributeSource in org.apache.lucene.analysis.el |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.el | |
---|---|
class |
GreekLowerCaseFilter
Normalizes token text to lower case, removes some Greek diacritics, and standardizes final sigma to sigma. |
class |
GreekStemFilter
A TokenFilter that applies GreekStemmer to stem Greek
words. |
Uses of AttributeSource in org.apache.lucene.analysis.en |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.en | |
---|---|
class |
EnglishMinimalStemFilter
A TokenFilter that applies EnglishMinimalStemmer to stem
English words. |
class |
EnglishPossessiveFilter
TokenFilter that removes possessives (trailing 's) from words. |
Uses of AttributeSource in org.apache.lucene.analysis.es |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.es | |
---|---|
class |
SpanishLightStemFilter
A TokenFilter that applies SpanishLightStemmer to stem Spanish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.fa |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fa | |
---|---|
class |
PersianNormalizationFilter
A TokenFilter that applies PersianNormalizer to normalize the
orthography. |
Uses of AttributeSource in org.apache.lucene.analysis.fi |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fi | |
---|---|
class |
FinnishLightStemFilter
A TokenFilter that applies FinnishLightStemmer to stem Finnish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.fr |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fr | |
---|---|
class |
ElisionFilter
Removes elisions from a TokenStream . |
class |
FrenchLightStemFilter
A TokenFilter that applies FrenchLightStemmer to stem French
words. |
class |
FrenchMinimalStemFilter
A TokenFilter that applies FrenchMinimalStemmer to stem French
words. |
class |
FrenchStemFilter
Deprecated. Use SnowballFilter with
FrenchStemmer instead, which has the
same functionality. This filter will be removed in Lucene 5.0 |
Uses of AttributeSource in org.apache.lucene.analysis.gl |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.gl | |
---|---|
class |
GalicianStemFilter
A TokenFilter that applies GalicianStemmer to stem
Galician words. |
Uses of AttributeSource in org.apache.lucene.analysis.hi |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.hi | |
---|---|
class |
HindiNormalizationFilter
A TokenFilter that applies HindiNormalizer to normalize the
orthography. |
class |
HindiStemFilter
A TokenFilter that applies HindiStemmer to stem Hindi words. |
Uses of AttributeSource in org.apache.lucene.analysis.hu |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.hu | |
---|---|
class |
HungarianLightStemFilter
A TokenFilter that applies HungarianLightStemmer to stem
Hungarian words. |
Uses of AttributeSource in org.apache.lucene.analysis.icu |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.icu | |
---|---|
class |
ICUFoldingFilter
A TokenFilter that applies search term folding to Unicode text, applying foldings from UTR#30 Character Foldings. |
class |
ICUNormalizer2Filter
Normalize token text with ICU's Normalizer2 |
class |
ICUTransformFilter
A TokenFilter that transforms text with ICU. |
Uses of AttributeSource in org.apache.lucene.analysis.icu.segmentation |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.icu.segmentation | |
---|---|
class |
ICUTokenizer
Breaks text into words according to UAX #29: Unicode Text Segmentation (http://www.unicode.org/reports/tr29/) |
Uses of AttributeSource in org.apache.lucene.analysis.id |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.id | |
---|---|
class |
IndonesianStemFilter
A TokenFilter that applies IndonesianStemmer to stem Indonesian words. |
Uses of AttributeSource in org.apache.lucene.analysis.in |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.in | |
---|---|
class |
IndicNormalizationFilter
A TokenFilter that applies IndicNormalizer to normalize text
in Indian Languages. |
class |
IndicTokenizer
Simple Tokenizer for text in Indian Languages. |
Constructors in org.apache.lucene.analysis.in with parameters of type AttributeSource | |
---|---|
IndicTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
|
Uses of AttributeSource in org.apache.lucene.analysis.it |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.it | |
---|---|
class |
ItalianLightStemFilter
A TokenFilter that applies ItalianLightStemmer to stem Italian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.miscellaneous |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.miscellaneous | |
---|---|
class |
EmptyTokenStream
An always exhausted token stream. |
class |
PrefixAndSuffixAwareTokenFilter
Links two PrefixAwareTokenFilter . |
class |
PrefixAwareTokenFilter
Joins two token streams and leaves the last token of the first stream available to be used when updating the token values in the second stream based on that token. |
class |
SingleTokenTokenStream
A TokenStream containing a single token. |
class |
StemmerOverrideFilter
Provides the ability to override any KeywordAttribute aware stemmer
with custom dictionary-based stemming. |
Uses of AttributeSource in org.apache.lucene.analysis.ngram |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ngram | |
---|---|
class |
EdgeNGramTokenFilter
Tokenizes the given token into n-grams of given size(s). |
class |
EdgeNGramTokenizer
Tokenizes the input from an edge into n-grams of given size(s). |
class |
NGramTokenFilter
Tokenizes the input into n-grams of the given size(s). |
class |
NGramTokenizer
Tokenizes the input into n-grams of the given size(s). |
Constructors in org.apache.lucene.analysis.ngram with parameters of type AttributeSource | |
---|---|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
EdgeNGramTokenizer.Side side,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
String sideLabel,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
NGramTokenizer(AttributeSource source,
Reader input,
int minGram,
int maxGram)
Creates NGramTokenizer with given min and max n-grams. |
Uses of AttributeSource in org.apache.lucene.analysis.nl |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.nl | |
---|---|
class |
DutchStemFilter
Deprecated. Use SnowballFilter with
DutchStemmer instead, which has the
same functionality. This filter will be removed in Lucene 5.0 |
Uses of AttributeSource in org.apache.lucene.analysis.path |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.path | |
---|---|
class |
PathHierarchyTokenizer
Take something like: |
Uses of AttributeSource in org.apache.lucene.analysis.payloads |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.payloads | |
---|---|
class |
DelimitedPayloadTokenFilter
Characters before the delimiter are the "token", those after are the payload. |
class |
NumericPayloadTokenFilter
Assigns a payload to a token based on the Token.type() |
class |
TokenOffsetPayloadTokenFilter
Adds the Token.setStartOffset(int)
and Token.setEndOffset(int)
First 4 bytes are the start |
class |
TypeAsPayloadTokenFilter
Makes the Token.type() a payload. |
Uses of AttributeSource in org.apache.lucene.analysis.position |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.position | |
---|---|
class |
PositionFilter
Set the positionIncrement of all tokens to the "positionIncrement", except the first return token which retains its original positionIncrement value. |
Uses of AttributeSource in org.apache.lucene.analysis.pt |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.pt | |
---|---|
class |
PortugueseLightStemFilter
A TokenFilter that applies PortugueseLightStemmer to stem
Portuguese words. |
class |
PortugueseMinimalStemFilter
A TokenFilter that applies PortugueseMinimalStemmer to stem
Portuguese words. |
class |
PortugueseStemFilter
A TokenFilter that applies PortugueseStemmer to stem
Portuguese words. |
Uses of AttributeSource in org.apache.lucene.analysis.reverse |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.reverse | |
---|---|
class |
ReverseStringFilter
Reverse token string, for example "country" => "yrtnuoc". |
Uses of AttributeSource in org.apache.lucene.analysis.ru |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ru | |
---|---|
class |
RussianLetterTokenizer
Deprecated. Use StandardTokenizer instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
class |
RussianLightStemFilter
A TokenFilter that applies RussianLightStemmer to stem Russian
words. |
class |
RussianLowerCaseFilter
Deprecated. Use LowerCaseFilter instead, which has the same
functionality. This filter will be removed in Lucene 4.0 |
class |
RussianStemFilter
Deprecated. Use SnowballFilter with
RussianStemmer instead, which has the
same functionality. This filter will be removed in Lucene 4.0 |
Constructors in org.apache.lucene.analysis.ru with parameters of type AttributeSource | |
---|---|
RussianLetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use RussianLetterTokenizer.RussianLetterTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
RussianLetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Deprecated. Construct a new RussianLetterTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.shingle |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.shingle | |
---|---|
class |
ShingleFilter
A ShingleFilter constructs shingles (token n-grams) from a token stream. |
class |
ShingleMatrixFilter
Deprecated. Will be removed in Lucene 4.0. This filter is unmaintained and might not behave correctly if used with custom Attributes, i.e. Attributes other than the ones located in org.apache.lucene.analysis.tokenattributes . It also uses
hardcoded payload encoders which makes it not easily adaptable to other use-cases. |
Uses of AttributeSource in org.apache.lucene.analysis.sinks |
---|
Methods in org.apache.lucene.analysis.sinks with parameters of type AttributeSource | |
---|---|
boolean |
TokenTypeSinkFilter.accept(AttributeSource source)
|
boolean |
TokenRangeSinkFilter.accept(AttributeSource source)
|
boolean |
DateRecognizerSinkFilter.accept(AttributeSource source)
|
Uses of AttributeSource in org.apache.lucene.analysis.snowball |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.snowball | |
---|---|
class |
SnowballFilter
A filter that stems words using a Snowball-generated stemmer. |
Uses of AttributeSource in org.apache.lucene.analysis.standard |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.standard | |
---|---|
class |
ClassicFilter
Normalizes tokens extracted with ClassicTokenizer . |
class |
ClassicTokenizer
A grammar-based tokenizer constructed with JFlex |
class |
StandardFilter
Normalizes tokens extracted with StandardTokenizer . |
class |
StandardTokenizer
A grammar-based tokenizer constructed with JFlex. |
class |
UAX29URLEmailTokenizer
This class implements Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29 URLs and email addresses are also tokenized according to the relevant RFCs. |
Constructors in org.apache.lucene.analysis.standard with parameters of type AttributeSource | |
---|---|
ClassicTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new ClassicTokenizer with a given AttributeSource . |
|
StandardTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new StandardTokenizer with a given AttributeSource . |
|
UAX29URLEmailTokenizer(AttributeSource source,
Reader input)
|
Uses of AttributeSource in org.apache.lucene.analysis.stempel |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.stempel | |
---|---|
class |
StempelFilter
Transforms the token stream as per the stemming algorithm. |
Uses of AttributeSource in org.apache.lucene.analysis.sv |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.sv | |
---|---|
class |
SwedishLightStemFilter
A TokenFilter that applies SwedishLightStemmer to stem Swedish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.th |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.th | |
---|---|
class |
ThaiWordFilter
TokenFilter that use BreakIterator to break each
Token that is Thai into separate Token(s) for each Thai word. |
Uses of AttributeSource in org.apache.lucene.analysis.tr |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.tr | |
---|---|
class |
TurkishLowerCaseFilter
Normalizes Turkish token text to lower case. |
Uses of AttributeSource in org.apache.lucene.analysis.wikipedia |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.wikipedia | |
---|---|
class |
WikipediaTokenizer
Extension of StandardTokenizer that is aware of Wikipedia syntax. |
Constructors in org.apache.lucene.analysis.wikipedia with parameters of type AttributeSource | |
---|---|
WikipediaTokenizer(AttributeSource source,
Reader input,
int tokenOutput,
Set<String> untokenizedTypes)
Creates a new instance of the WikipediaTokenizer . |
Uses of AttributeSource in org.apache.lucene.collation |
---|
Subclasses of AttributeSource in org.apache.lucene.collation | |
---|---|
class |
CollationKeyFilter
Converts each token into its CollationKey , and then
encodes the CollationKey with IndexableBinaryStringTools , to allow
it to be stored as an index term. |
class |
ICUCollationKeyFilter
Converts each token into its CollationKey , and
then encodes the CollationKey with IndexableBinaryStringTools , to
allow it to be stored as an index term. |
Uses of AttributeSource in org.apache.lucene.index |
---|
Methods in org.apache.lucene.index that return AttributeSource | |
---|---|
AttributeSource |
FieldInvertState.getAttributeSource()
|
Uses of AttributeSource in org.apache.lucene.queryParser.core.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.core.config | |
---|---|
class |
FieldConfig
This class represents a field configuration. |
class |
QueryConfigHandler
This class can be used to hold any query configuration and no field configuration. |
Uses of AttributeSource in org.apache.lucene.queryParser.standard.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.standard.config | |
---|---|
class |
StandardQueryConfigHandler
This query configuration handler is used for almost every processor defined in the StandardQueryNodeProcessorPipeline processor pipeline. |
Uses of AttributeSource in org.apache.lucene.search.highlight |
---|
Subclasses of AttributeSource in org.apache.lucene.search.highlight | |
---|---|
class |
TokenStreamFromTermPositionVector
|
Uses of AttributeSource in org.apache.lucene.util |
---|
Methods in org.apache.lucene.util that return AttributeSource | |
---|---|
AttributeSource |
AttributeSource.cloneAttributes()
Performs a clone of all AttributeImpl instances returned in a new
AttributeSource instance. |
Methods in org.apache.lucene.util with parameters of type AttributeSource | |
---|---|
void |
AttributeSource.copyTo(AttributeSource target)
Copies the contents of this AttributeSource to the given target AttributeSource . |
Constructors in org.apache.lucene.util with parameters of type AttributeSource | |
---|---|
AttributeSource(AttributeSource input)
An AttributeSource that uses the same attributes as the supplied one. |
Uses of AttributeSource in org.apache.lucene.wordnet |
---|
Subclasses of AttributeSource in org.apache.lucene.wordnet | |
---|---|
class |
SynonymTokenFilter
Injects additional tokens for synonyms of token terms fetched from the underlying child stream; the child stream must deliver lowercase tokens for synonyms to be found. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |