Package | Description |
---|---|
org.crosswire.jsword.index.lucene.analysis |
Implementation of various Lucene analyzers, providing language dependent customizations.
|
Modifier and Type | Class and Description |
---|---|
class |
ArabicLuceneAnalyzer
An Analyzer whose
TokenStream is built from a
ArabicLetterTokenizer filtered with LowerCaseFilter ,
ArabicNormalizationFilter , ArabicStemFilter (optional) and
Arabic StopFilter (optional). |
class |
ChineseLuceneAnalyzer
Uses org.apache.lucene.analysis.cn.ChineseAnalyzer Analysis:
ChineseTokenizer, ChineseFilter StopFilter, Stemming not implemented yet
Note: org.apache.lucene.analysis.cn.CJKAnalyzer takes overlapping two
character tokenization approach which leads to larger index size.
|
class |
ConfigurableSnowballAnalyzer
An Analyzer whose
TokenStream is built from a
LowerCaseTokenizer filtered with SnowballFilter (optional)
and StopFilter (optional) Default behavior: Stemming is done, Stop
words not removed A snowball stemmer is configured according to the language
of the Book. |
class |
CzechLuceneAnalyzer
An Analyzer whose
TokenStream is built from a
LowerCaseTokenizer filtered with StopFilter (optional). |
class |
EnglishLuceneAnalyzer
English Analyzer works like lucene SimpleAnalyzer + Stemming.
|
class |
GermanLuceneAnalyzer
Based on Lucene's GermanAnalyzer
|
class |
GreekLuceneAnalyzer
Uses org.apache.lucene.analysis.el.GreekAnalyzer to do lowercasing and
stopword(off by default).
|
class |
KeyAnalyzer
A specialized analyzer that normalizes Strong's Numbers.
|
class |
MorphologyAnalyzer
Robinson Morphological Codes are separated by whitespace.
|
class |
PersianLuceneAnalyzer
An Analyzer whose
TokenStream is built from a
ArabicLetterTokenizer filtered with LowerCaseFilter ,
ArabicNormalizationFilter , PersianNormalizationFilter and
Persian StopFilter (optional) |
class |
SimpleLuceneAnalyzer
Simple Analyzer providing same function as
org.apache.lucene.analysis.SimpleAnalyzer This is intended to be the default
analyzer for natural language fields.
|
class |
SmartChineseLuceneAnalyzer
A simple wrapper for
SmartChineseAnalyzer , which takes overlapping
two character tokenization approach which leads to larger index size, like
org.apache.lucene.analyzer.cjk.CJKAnalyzer . |
class |
StrongsNumberAnalyzer
A specialized analyzer that normalizes JSword keys.
|
class |
ThaiLuceneAnalyzer
Tokenization using ThaiWordFilter.
|
class |
XRefAnalyzer
A specialized analyzer that normalizes Cross References.
|
Modifier and Type | Method and Description |
---|---|
AbstractBookAnalyzer |
AnalyzerFactory.createAnalyzer(Book book) |