|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use AttributeSource | |
---|---|
org.apache.lucene.analysis | API and code to convert text into indexable/searchable tokens. |
org.apache.lucene.analysis.ar | Analyzer for Arabic. |
org.apache.lucene.analysis.br | Analyzer for Brazilian Portuguese. |
org.apache.lucene.analysis.cjk | Analyzer for Chinese, Japanese, and Korean, which indexes bigrams (overlapping groups of two adjacent Han characters). |
org.apache.lucene.analysis.cn | Analyzer for Chinese, which indexes unigrams (individual chinese characters). |
org.apache.lucene.analysis.cn.smart |
Analyzer for Simplified Chinese, which indexes words. |
org.apache.lucene.analysis.compound | A filter that decomposes compound words you find in many Germanic languages into the word parts. |
org.apache.lucene.analysis.de | Analyzer for German. |
org.apache.lucene.analysis.el | Analyzer for Greek. |
org.apache.lucene.analysis.fa | Analyzer for Persian. |
org.apache.lucene.analysis.fr | Analyzer for French. |
org.apache.lucene.analysis.miscellaneous | Miscellaneous TokenStreams |
org.apache.lucene.analysis.ngram | Character n-gram tokenizers and filters. |
org.apache.lucene.analysis.nl | Analyzer for Dutch. |
org.apache.lucene.analysis.payloads | Provides various convenience classes for creating payloads on Tokens. |
org.apache.lucene.analysis.position | Filter for assigning position increments. |
org.apache.lucene.analysis.reverse | Filter to reverse token text. |
org.apache.lucene.analysis.ru | Analyzer for Russian. |
org.apache.lucene.analysis.shingle | Word n-gram filters |
org.apache.lucene.analysis.sinks | Implementations of the SinkTokenizer that might be useful. |
org.apache.lucene.analysis.snowball | TokenFilter and Analyzer implementations that use Snowball
stemmers. |
org.apache.lucene.analysis.standard | A fast grammar-based tokenizer constructed with JFlex. |
org.apache.lucene.analysis.th | Analyzer for Thai. |
org.apache.lucene.collation |
CollationKeyFilter and ICUCollationKeyFilter
convert each token into its binary CollationKey using the
provided Collator , and then encode the CollationKey
as a String using
IndexableBinaryStringTools , to allow it to be
stored as an index term. |
org.apache.lucene.index | Code to maintain and access indices. |
org.apache.lucene.queryParser.core.config | Contains the base classes used to configure the query processing |
org.apache.lucene.queryParser.standard.config | Standard Lucene Query Configuration |
org.apache.lucene.util | Some utility classes. |
org.apache.lucene.wikipedia.analysis | Tokenizer that is aware of Wikipedia syntax. |
org.apache.lucene.wordnet | This package uses synonyms defined by WordNet. |
Uses of AttributeSource in org.apache.lucene.analysis |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis | |
---|---|
class |
ASCIIFoldingFilter
This class converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if one exists. |
class |
CachingTokenFilter
This class can be used if the token attributes of a TokenStream are intended to be consumed more than once. |
class |
CharTokenizer
An abstract base class for simple, character-oriented tokenizers. |
class |
ISOLatin1AccentFilter
Deprecated. If you build a new index, use ASCIIFoldingFilter
which covers a superset of Latin 1.
This class is included for use with existing
indexes and will be removed in a future release (possibly Lucene 4.0). |
class |
KeywordTokenizer
Emits the entire input as a single token. |
class |
LengthFilter
Removes words that are too long or too short from the stream. |
class |
LetterTokenizer
A LetterTokenizer is a tokenizer that divides text at non-letters. |
class |
LowerCaseFilter
Normalizes token text to lower case. |
class |
LowerCaseTokenizer
LowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together. |
class |
NumericTokenStream
Expert: This class provides a TokenStream
for indexing numeric values that can be used by NumericRangeQuery or NumericRangeFilter . |
class |
PorterStemFilter
Transforms the token stream as per the Porter stemming algorithm. |
class |
StopFilter
Removes stop words from a token stream. |
class |
TeeSinkTokenFilter
This TokenFilter provides the ability to set aside attribute states that have already been analyzed. |
static class |
TeeSinkTokenFilter.SinkTokenStream
|
class |
TokenFilter
A TokenFilter is a TokenStream whose input is another TokenStream. |
class |
Tokenizer
A Tokenizer is a TokenStream whose input is a Reader. |
class |
TokenStream
A TokenStream enumerates the sequence of tokens, either from
Field s of a Document or from query text. |
class |
WhitespaceTokenizer
A WhitespaceTokenizer is a tokenizer that divides text at whitespace. |
Methods in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
abstract boolean |
TeeSinkTokenFilter.SinkFilter.accept(AttributeSource source)
Returns true, iff the current state of the passed-in AttributeSource shall be stored
in the sink. |
Constructors in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
CharTokenizer(AttributeSource source,
Reader input)
|
|
KeywordTokenizer(AttributeSource source,
Reader input,
int bufferSize)
|
|
LetterTokenizer(AttributeSource source,
Reader in)
Construct a new LetterTokenizer using a given AttributeSource . |
|
LowerCaseTokenizer(AttributeSource source,
Reader in)
Construct a new LowerCaseTokenizer using a given AttributeSource . |
|
NumericTokenStream(AttributeSource source,
int precisionStep)
Expert: Creates a token stream for numeric values with the specified precisionStep using the given AttributeSource . |
|
Tokenizer(AttributeSource source)
Construct a token stream processing the given input using the given AttributeSource. |
|
Tokenizer(AttributeSource source,
Reader input)
Construct a token stream processing the given input using the given AttributeSource. |
|
TokenStream(AttributeSource input)
A TokenStream that uses the same attributes as the supplied one. |
|
WhitespaceTokenizer(AttributeSource source,
Reader in)
Construct a new WhitespaceTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.ar |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ar | |
---|---|
class |
ArabicLetterTokenizer
Tokenizer that breaks text into runs of letters and diacritics. |
class |
ArabicNormalizationFilter
A TokenFilter that applies ArabicNormalizer to normalize the orthography. |
class |
ArabicStemFilter
A TokenFilter that applies ArabicStemmer to stem Arabic words.. |
Constructors in org.apache.lucene.analysis.ar with parameters of type AttributeSource | |
---|---|
ArabicLetterTokenizer(AttributeSource source,
Reader in)
|
Uses of AttributeSource in org.apache.lucene.analysis.br |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.br | |
---|---|
class |
BrazilianStemFilter
A TokenFilter that applies BrazilianStemmer . |
Uses of AttributeSource in org.apache.lucene.analysis.cjk |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cjk | |
---|---|
class |
CJKTokenizer
CJKTokenizer is designed for Chinese, Japanese, and Korean languages. |
Constructors in org.apache.lucene.analysis.cjk with parameters of type AttributeSource | |
---|---|
CJKTokenizer(AttributeSource source,
Reader in)
|
Uses of AttributeSource in org.apache.lucene.analysis.cn |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn | |
---|---|
class |
ChineseFilter
A TokenFilter with a stop word table. |
class |
ChineseTokenizer
Tokenize Chinese text as individual chinese characters. |
Constructors in org.apache.lucene.analysis.cn with parameters of type AttributeSource | |
---|---|
ChineseTokenizer(AttributeSource source,
Reader in)
|
Uses of AttributeSource in org.apache.lucene.analysis.cn.smart |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn.smart | |
---|---|
class |
SentenceTokenizer
Tokenizes input text into sentences. |
class |
WordTokenFilter
A TokenFilter that breaks sentences into words. |
Constructors in org.apache.lucene.analysis.cn.smart with parameters of type AttributeSource | |
---|---|
SentenceTokenizer(AttributeSource source,
Reader reader)
|
Uses of AttributeSource in org.apache.lucene.analysis.compound |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.compound | |
---|---|
class |
CompoundWordTokenFilterBase
Base class for decomposition token filters. |
class |
DictionaryCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
class |
HyphenationCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
Uses of AttributeSource in org.apache.lucene.analysis.de |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.de | |
---|---|
class |
GermanStemFilter
A TokenFilter that stems German words. |
Uses of AttributeSource in org.apache.lucene.analysis.el |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.el | |
---|---|
class |
GreekLowerCaseFilter
Normalizes token text to lower case, removes some Greek diacritics, and standardizes final sigma to sigma. |
Uses of AttributeSource in org.apache.lucene.analysis.fa |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fa | |
---|---|
class |
PersianNormalizationFilter
A TokenFilter that applies PersianNormalizer to normalize the
orthography. |
Uses of AttributeSource in org.apache.lucene.analysis.fr |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fr | |
---|---|
class |
ElisionFilter
Removes elisions from a TokenStream . |
class |
FrenchStemFilter
A TokenFilter that stems french words. |
Uses of AttributeSource in org.apache.lucene.analysis.miscellaneous |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.miscellaneous | |
---|---|
class |
EmptyTokenStream
An always exhausted token stream. |
class |
PrefixAndSuffixAwareTokenFilter
Links two PrefixAwareTokenFilter . |
class |
PrefixAwareTokenFilter
Joins two token streams and leaves the last token of the first stream available to be used when updating the token values in the second stream based on that token. |
class |
SingleTokenTokenStream
A TokenStream containing a single token. |
Uses of AttributeSource in org.apache.lucene.analysis.ngram |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ngram | |
---|---|
class |
EdgeNGramTokenFilter
Tokenizes the given token into n-grams of given size(s). |
class |
EdgeNGramTokenizer
Tokenizes the input from an edge into n-grams of given size(s). |
class |
NGramTokenFilter
Tokenizes the input into n-grams of the given size(s). |
class |
NGramTokenizer
Tokenizes the input into n-grams of the given size(s). |
Constructors in org.apache.lucene.analysis.ngram with parameters of type AttributeSource | |
---|---|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
EdgeNGramTokenizer.Side side,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
String sideLabel,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
NGramTokenizer(AttributeSource source,
Reader input,
int minGram,
int maxGram)
Creates NGramTokenizer with given min and max n-grams. |
Uses of AttributeSource in org.apache.lucene.analysis.nl |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.nl | |
---|---|
class |
DutchStemFilter
A TokenFilter that stems Dutch words. |
Uses of AttributeSource in org.apache.lucene.analysis.payloads |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.payloads | |
---|---|
class |
DelimitedPayloadTokenFilter
Characters before the delimiter are the "token", those after are the payload. |
class |
NumericPayloadTokenFilter
Assigns a payload to a token based on the Token.type() |
class |
TokenOffsetPayloadTokenFilter
Adds the Token.setStartOffset(int)
and Token.setEndOffset(int)
First 4 bytes are the start |
class |
TypeAsPayloadTokenFilter
Makes the Token.type() a payload. |
Uses of AttributeSource in org.apache.lucene.analysis.position |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.position | |
---|---|
class |
PositionFilter
Set the positionIncrement of all tokens to the "positionIncrement", except the first return token which retains its original positionIncrement value. |
Uses of AttributeSource in org.apache.lucene.analysis.reverse |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.reverse | |
---|---|
class |
ReverseStringFilter
Reverse token string, for example "country" => "yrtnuoc". |
Uses of AttributeSource in org.apache.lucene.analysis.ru |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ru | |
---|---|
class |
RussianLetterTokenizer
A RussianLetterTokenizer is a Tokenizer that extends LetterTokenizer
by also allowing the basic latin digits 0-9. |
class |
RussianLowerCaseFilter
Deprecated. Use LowerCaseFilter instead, which has the same
functionality. This filter will be removed in Lucene 4.0 |
class |
RussianStemFilter
A TokenFilter that stems Russian words. |
Constructors in org.apache.lucene.analysis.ru with parameters of type AttributeSource | |
---|---|
RussianLetterTokenizer(AttributeSource source,
Reader in)
|
Uses of AttributeSource in org.apache.lucene.analysis.shingle |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.shingle | |
---|---|
class |
ShingleFilter
A ShingleFilter constructs shingles (token n-grams) from a token stream. |
class |
ShingleMatrixFilter
A ShingleMatrixFilter constructs shingles (token n-grams) from a token stream. |
Uses of AttributeSource in org.apache.lucene.analysis.sinks |
---|
Methods in org.apache.lucene.analysis.sinks with parameters of type AttributeSource | |
---|---|
boolean |
TokenTypeSinkFilter.accept(AttributeSource source)
|
boolean |
TokenRangeSinkFilter.accept(AttributeSource source)
|
boolean |
DateRecognizerSinkFilter.accept(AttributeSource source)
|
Uses of AttributeSource in org.apache.lucene.analysis.snowball |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.snowball | |
---|---|
class |
SnowballFilter
A filter that stems words using a Snowball-generated stemmer. |
Uses of AttributeSource in org.apache.lucene.analysis.standard |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.standard | |
---|---|
class |
StandardFilter
Normalizes tokens extracted with StandardTokenizer . |
class |
StandardTokenizer
A grammar-based tokenizer constructed with JFlex |
Constructors in org.apache.lucene.analysis.standard with parameters of type AttributeSource | |
---|---|
StandardTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new StandardTokenizer with a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.th |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.th | |
---|---|
class |
ThaiWordFilter
TokenFilter that use BreakIterator to break each
Token that is Thai into separate Token(s) for each Thai word. |
Uses of AttributeSource in org.apache.lucene.collation |
---|
Subclasses of AttributeSource in org.apache.lucene.collation | |
---|---|
class |
CollationKeyFilter
Converts each token into its CollationKey , and then
encodes the CollationKey with IndexableBinaryStringTools , to allow
it to be stored as an index term. |
class |
ICUCollationKeyFilter
Converts each token into its CollationKey , and
then encodes the CollationKey with IndexableBinaryStringTools , to
allow it to be stored as an index term. |
Uses of AttributeSource in org.apache.lucene.index |
---|
Methods in org.apache.lucene.index that return AttributeSource | |
---|---|
AttributeSource |
FieldInvertState.getAttributeSource()
|
Uses of AttributeSource in org.apache.lucene.queryParser.core.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.core.config | |
---|---|
class |
FieldConfig
This class represents a field configuration. |
class |
QueryConfigHandler
This class can be used to hold any query configuration and no field configuration. |
Uses of AttributeSource in org.apache.lucene.queryParser.standard.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.standard.config | |
---|---|
class |
StandardQueryConfigHandler
This query configuration handler is used for almost every processor defined in the StandardQueryNodeProcessorPipeline processor pipeline. |
Uses of AttributeSource in org.apache.lucene.util |
---|
Methods in org.apache.lucene.util that return AttributeSource | |
---|---|
AttributeSource |
AttributeSource.cloneAttributes()
Performs a clone of all AttributeImpl instances returned in a new
AttributeSource instance. |
Constructors in org.apache.lucene.util with parameters of type AttributeSource | |
---|---|
AttributeSource(AttributeSource input)
An AttributeSource that uses the same attributes as the supplied one. |
Uses of AttributeSource in org.apache.lucene.wikipedia.analysis |
---|
Subclasses of AttributeSource in org.apache.lucene.wikipedia.analysis | |
---|---|
class |
WikipediaTokenizer
Extension of StandardTokenizer that is aware of Wikipedia syntax. |
Constructors in org.apache.lucene.wikipedia.analysis with parameters of type AttributeSource | |
---|---|
WikipediaTokenizer(AttributeSource source,
Reader input,
int tokenOutput,
Set<String> untokenizedTypes)
Creates a new instance of the WikipediaTokenizer . |
Uses of AttributeSource in org.apache.lucene.wordnet |
---|
Subclasses of AttributeSource in org.apache.lucene.wordnet | |
---|---|
class |
SynonymTokenFilter
Injects additional tokens for synonyms of token terms fetched from the underlying child stream; the child stream must deliver lowercase tokens for synonyms to be found. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |