Get the n-gram from the end of the input
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Backs up the input stream by amount steps.
Implimented as:
(x <= min) ? base : sqrt(x+(base**2)-min)
...but with a special case check for 0.
Create documents for the test.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginColumn - field in class org.apache.lucene.queryParser.surround.parser.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine - field in class org.apache.lucene.queryParser.precedence.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine - field in class org.apache.lucene.queryParser.surround.parser.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Returns the next character that marks the beginning of the next token.
Run the benchmark algorithm.
Benchmark according to the implementation, using the workingDir as the place to store things.
Benchmark according to the implementation, using the workingDir as the place to store things.
Marker Interface defining some common options.
The value of the field in Binary, or null.
The value of the field in Binary, or null.
Returns a BitSet with true for documents which should be permitted in
search results, and false for those that should not.
Returns a BitSet with true for documents which should be permitted in
search results, and false for those that should not.
Returns a BitSet with true for documents which should be permitted in
search results, and false for those that should not.
Returns a BitSet with true for documents which should be permitted in
search results, and false for those that should not.
Returns a BitSet with true for documents which should be
permitted in search results, and false for those that should
not.
Optimized implementation of a vector of bits.
Constructs a vector capable of holding n
bits.
A clause in a BooleanQuery.
Constructs a BooleanClause.
A Query that matches documents matching boolean combinations of other
queries, e.g.
Constructs an empty boolean query.
Constructs an empty boolean query.
List of typical Brazilian stopwords.
Analyzer for Brazilian language.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Based on GermanStemFilter
Construct a token stream filtering the given input.
A stemmer for Brazilian words.
Base implementation class for buffered
IndexInput
.
byteToFloat(b, mantissaBits=3, zeroExponent=15)
byteToFloat(b, mantissaBits=5, zeroExponent=2)
Converts an 8 bit float to a 32 bit float.
Wraps another filter's result and caches it.
Allows multiple
Filter
s to be chained.
This interface describes a character stream that maintains line and
column number positions of the characters.
This interface describes a character stream that maintains line and
column number positions of the characters.
This interface describes a character stream that maintains line and
column number positions of the characters.
This interface describes a character stream that maintains line and
column number positions of the characters.
An abstract base class for simple, character-oriented tokenizers.
Title: ChineseAnalyzer
Description:
Subclass of org.apache.lucene.analysis.Analyzer
build from a ChineseTokenizer, filtered with ChineseFilter.
Title: ChineseFilter
Description: Filter with a stop word table
Rule: No digital is allowed.
Title: ChineseTokenizer
Description: Extract tokens from the Stream using Character.getType()
Rule: A Chinese character as a single token
Copyright: Copyright (c) 2001
Company:
The difference between thr ChineseTokenizer and the
CJKTokenizer (id=23545) is that they have different
token parsing logic.
Construct a token stream processing the given input.
Filters CJKTokenizer with StopFilter.
Builds an analyzer which removes words in
STOP_WORDS
.
Builds an analyzer which removes words in the provided array.
CJKTokenizer was modified from StopTokenizer which does a decent job for
most European languages.
Construct a token stream processing the given input.
Returns the list of clauses in this query.
Removes all entries from the PriorityQueue.
Removes all entries from the ScorerDocQueue.
Sets the value of bit
to zero.
Clear all data, prepare for more tests.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Attempt to clear (forcefully unlock and remove) the
specified lock.
Returns a clone of this stream.
Create a shallow copy of us -- used in rewriting if necessary
Returns a clone of this stream.
clone() - method in class org.apache.lucene.benchmark.byTask.tasks.
PerfTask Returns a clone of this query.
Returns a copy of this instance, also copying each
SegmentInfo.
clone() - method in class org.apache.lucene.queryParser.surround.query.
SrndQuery Closes this stream to further operations.
Closes the enumeration to further activity, freeing resources.
Frees associated resources.
Closes the enumeration to further activity, freeing resources.
Closes the store to future operations.
Closes the enumeration to further activity, freeing resources.
Closes the stream to futher operations.
Close this index, writing all pending changes to disk.
Closes this stream to further operations.
Closes files associated with this index.
Note that the underlying IndexReader is not closed, if
IndexSearcher was constructed with IndexSearcher(IndexReader r).
Flushes all changes to an index and closes all
associated files.
Closes the store to future operations, releasing associated memory.
Closes this stream to further operations.
Closes the enumeration to further activity, freeing resources.
Frees resources associated with this Searcher.
Frees resources associated with this Searcher.
Frees associated resources.
Closes the enumeration to further activity, freeing resources.
Close the input TokenStream.
By default, closes the input Reader.
Releases resources associated with this stream.
Closes the enumeration to further activity, freeing resources.
Called once for every non-zero scoring document, with the document number
and its score.
Called once for every non-zero scoring document, with the document number
and its score.
Called once for every non-zero scoring document, with the document number
and its score.
Expert: called when re-writing queries under MultiSearcher.
Commit changes resulting from delete, undeleteAll, or
setNorm operations
If an exception is hit, then either no changes or all
changes will have been committed to the index
(transactional semantics).
Stores a comparator corresponding to each field being sorted by
Compares two ScoreDoc objects and returns a result indicating their
sort order.
Compares two terms, returning a negative integer if this
term belongs before the argument, zero if this term is equal to the
argument, and a positive integer if this term belongs after the argument.
Called by the constructor of
RegexTermEnum
allowing
implementations to cache a compiled version of the regular
expression pattern.
Expert: Describes the score computation for document and query, andcan distinguish a match independent of a positive value.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Store the original field value in the index in a compressed form.
Config - class org.apache.lucene.benchmark.byTask.utils.
Config Perf run configuration properties.
Read config from file containing both algorithm and config properties.
Create config without algorithm - usefull for a programmatic perf test.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
A query that wraps a filter and simply returns a constant score equal to the
query boost for every document in the filter.
A range query that returns a constant score equal to its boost for
all documents in the range.
Implemented as overlap / maxOverlap
.
Computes a score factor based on the fraction of all query terms that a
document contains.
Computes a score factor based on the fraction of all query terms that a
document contains.
Copy contents of a directory src to a directory dest.
count - field in class org.apache.lucene.benchmark.stats.
TimeData Number of records processed.
Returns the total number of one bits in this vector.
Close the IndexWriter and open an IndexReader.
Close the IndexReader and open an IndexWriter.
Creates a new, empty file in the directory with the given name.
Creates a new, empty file in the directory with the given name.
Creates a new, empty file in the directory with the given name.
Parse the strings containing Lucene queries.
Creates and returns a searcher that can be used to execute arbitrary
Lucene queries and to collect the resulting query results as hits.
Optimized construction of new Terms by reusing same field as this Term
- avoids field.intern() overhead
Creates and returns a token for the given synonym of the current input
token; Override for custom (stateless or stateful) behaviour, if desired.
Create weight in multiple index scenario.
creates a weight for query
Expert: Constructs an appropriate Weight implementation for this query.
Returns a Weight that applies the filter to the enclosed query's Weight.
Expert: Constructs an appropriate Weight implementation for this query.
Expert: Constructs an appropriate Weight implementation for this query.
Expert: Constructs an appropriate Weight implementation for this query.
This is the last token that has been consumed successfully.
This is the last token that has been consumed successfully.
This is the last token that has been consumed successfully.
This is the last token that has been consumed successfully.
This is the last token that has been consumed successfully.
Sort using a custom Comparator.
List of typical stopwords.
Analyzer for Czech language.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Generated class implementing code defined by a snowball script.
Provides support for converting dates to strings and vice-versa.
Provides support for converting dates to strings and vice-versa.
Converts a Date to a string suitable for indexing.
Converts a Date to a string suitable for indexing.
Decodes a normalization factor stored in an index.
Logical operation when none is declared.
Expert: The cache used internally by sorting and range query classes.
A lower-casing word analyzer with English stop words (can be shared
freely across threads without harm); global per class loader.
Default value is Integer.MAX_VALUE
.
Simple
Encoder
implementation that does not modify the output
Expert: Default scoring implementation.
Delete a document by docid.
Deletes the document numbered docNum
.
Deletes the document numbered docNum
.
Deletes all documents containing term
.
Deletes all documents that have a given term
indexed.
Deletes the document(s) containing term
.
Deletes the document(s) containing any of the
terms.
Removes an existing file in the directory.
Removes an existing file in the directory.
Removes an existing file in the directory.
Deletes documents from an index that do not contain a term.
A simple interface representing a Dictionary
Equality measure on the term
Equality measure on the term
Equality measure on the term
Equality measure on the term
A Directory is a flat list of files.
Returns the directory this index resides in.
A query that generates the union of documents produced by its subqueries, and that scores each document with the maximum
score for that document as produced by any subquery, plus a tie breaking increment for any additional matching subqueries.
Creates a new DisjunctionMaxQuery
Creates a new empty DisjunctionMaxQuery.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
When distanceSubQueryNotAllowed() returns non null, the reason why the subquery
is not allowed as a distance subquery is returned.
When distanceSubQueryNotAllowed() returns non null, the reason why the subquery
is not allowed as a distance subquery is returned.
Code to execute with exclusive access.
Subclass must implement this.
Expert: A hit document's number.
DOC - static field in class org.apache.lucene.search.
SortField Sort by document number (index order).
Returns the current document number.
Returns the current document number matching the query.
doc() - method in class org.apache.lucene.search.
Scorer Returns the current document number matching the query.
doc() - method in class org.apache.lucene.search.spans.
Spans Returns the document number of the current match.
Returns the current document number.
Returns the stored fields of the nth document in this set.
Expert: Returns the stored fields of document i
.
Expert: Returns the stored fields of document i
.
Expert: Returns the stored fields of document i
.
Returns the number of documents currently in this index.
Returns the number of documents currently in this index.
Returns the docFreq of the current Term in the enumeration.
Returns the docFreq of the current Term in the enumeration.
Returns the docFreq of the current Term in the enumeration.
Returns the number of documents containing the term t
.
Expert: Returns the number of documents containing term
.
TODO: parallelize this one too
Expert: Returns the number of documents containing term
.
Expert: Returns the number of documents containing term
.
Expert: For each term in the terms array, calculates the number of
documents containing term
.
Expert: For each term in the terms array, calculates the number of
documents containing term
.
Create documents for the test.
Documents are the unit of indexing and search.
Constructs a new document with no fields.
Makes a document for a File.
Creates a Lucene
Document
from a
File
.
Makes a document for a File.
Returns the stored fields of the n
th
Document
in this index.
Allows a class to act as a Lucene document handler
Constructs with chained exception.
Implements deletion of the document numbered docNum
.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
Perform the task once (ignoring repetions specification)
Return number of work items done by this task.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
Done() - method in class org.apache.lucene.queryParser.surround.parser.
CharStream The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
The lexer calls this function to indicate that it is done with the stream
and hence implementations can free any resources held by this class.
Implements setNorm in subclass.
Implements actual undeleteAll() in subclass.
Creates a Driver using Digester
List of typical Dutch stopwords.
Analyzer for Dutch language.
Builds an analyzer with the default stop words (
DUTCH_STOP_WORDS
)
and a few default entries for the stem exclusion table.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
A filter that stems Dutch words.
Builds a DutchStemFilter that uses an exclusiontable.
Generated class implementing code defined by a snowball script.
A stemmer for Dutch words.
Tokenizes the input into n-grams of the given size.
Creates EdgeNGramTokenizer that can generate an n-gram of the given size.
Elapsed time in milliseconds.
Encodes a normalization factor for storage in an index.
end() - method in class org.apache.lucene.search.spans.
Spans Returns the end position of the current match.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
endColumn - field in class org.apache.lucene.queryParser.precedence.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
endColumn - field in class org.apache.lucene.queryParser.surround.parser.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
Indicates the end of the enumeration has been reached
Indicates the end of the enumeration has been reached
Indicates the end of the enumeration has been reached
Indicates the end of the enumeration has been reached
endLine - field in class org.apache.lucene.analysis.standard.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
endLine - field in class org.apache.lucene.queryParser.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
endLine - field in class org.apache.lucene.queryParser.precedence.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
endLine - field in class org.apache.lucene.queryParser.surround.parser.
Token beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.
Returns this Token's ending offset, one greater than the position of the
last character corresponding to this token in the source text.
Mark the ending time of this test run.
An array containing some common English words that are not usually useful
for searching.
Generated class implementing code defined by a snowball script.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
The end of line string for this machine.
The end of line string for this machine.
The end of line string for this machine.
The end of line string for this machine.
The end of line string for this machine.
Returns true iff o
is equal to this.
Returns true iff o
is equal to this.
Returns true if o
is equal to this.
Returns true if o
is equal to this.
Return true iff we represent the same query as o
Returns true iff o
is equal to this.
Returns true if o
is equal to this.
Indicates whether some other object is "equal to" this one.
Returns true iff o
is equal to this.
Returns true iff o
is equal to this.
Returns true if o
is equal to this.
Returns true iff o
is equal to this.
Returns true iff o
is equal to this.
Returns true iff o
is equal to this.
Returns true iff o
is equal to this.
Compares two terms, returning true iff they have the same
field and text.
Returns true iff o
is equal to this.
Returns a String where those characters that QueryParser
expects to be escaped are escaped by a preceding \
.
Returns a String where those characters that QueryParser
expects to be escaped are escaped by a preceding \
.
Check whether the word exists in the index.
Perform synonym expansion on a query.
Perform synonym expansion on a query.
Each entry in this array is an array of integers.
Each entry in this array is an array of integers.
Each entry in this array is an array of integers.
Each entry in this array is an array of integers.
Each entry in this array is an array of integers.
Explain the score of a document.
Returns an explanation of the score for a document.
An explanation of the score computation for the named document.
Returns an Explanation that describes how doc
scored against
query
.
Expert: low-level implementation method
Returns an Explanation that describes how doc
scored against
weight
.
Expert: low-level implementation method
Returns an Explanation that describes how doc
scored against
weight
.
Expert: low-level implementation method
Returns an Explanation that describes how doc
scored against
weight
.
Expert: Describes the score computation for document and query.
A lower-casing word analyzer with extended English stop words
(can be shared freely across threads without harm); global per class
loader.
Override if you wish to change what is extracted
Split the Reuters SGML documents into Simple Text files containing: Title, Date, Dateline, Body
Expert: adds all terms occuring in this query to the terms set.
Expert: adds all terms occuring in this query to the terms set.
Expert: adds all terms occuring in this query to the terms set.
Expert: adds all terms occuring in this query to the terms set.
Expert: adds all terms occuring in this query to the terms set.
Expert: adds all terms occuring in this query to the terms set.
Field name for each word in the ngram index.
If a lock obtain called, this failureReason may be set
with the "root cause" Exception as to why the lock was
not obtained.
An efficient implementation of JavaCC's CharStream interface.
An efficient implementation of JavaCC's CharStream interface.
An efficient implementation of JavaCC's CharStream interface.
An efficient implementation of JavaCC's CharStream interface.
Constructs from a Reader.
Constructs from a Reader.
Constructs from a Reader.
Constructs from a Reader.
A field is a section of a Document.
Returns the field of this term, an interned string.
Create a stored field with binary value.
Create a tokenized and indexed field that is not stored.
Create a tokenized and indexed field that is not stored, optionally with
storing term vectors.
Create a field by specifying its name, value and how it will
be saved in the index.
Create a field by specifying its name, value and how it will
be saved in the index.
Represents sorting by document number (index order).
Represents sorting by document score (relevancy).
Expert: Maintains caches of term values.
Expert: A ScoreDoc which also contains information about
how to sort the referenced document.
Expert: Creates one of these objects with empty sort information.
Expert: Creates one of these objects with the given sort information.
Given a directory and a list of fields, updates the fieldNorms in place for every document.
Constructor for code that wishes to use this class programatically
If Similarity is null, kill the field norms.
Constructs a new runtime exception with null
as its
detail message.
Constructs a new runtime exception with the specified detail message.
Constructs a new runtime exception with the specified detail message and
cause.
Constructs a new runtime exception with the specified cause and a
detail message of (cause==null ? null : cause.toString())
(which typically contains the class and detail message of
cause).
Expert: The values which are used to sort the referenced document.
Stores the sort criteria being used.
The fields which were used to sort results by.
Returns an Enumeration of all the fields in a document.
Provides information about what should be done with this Field
Expert: A hit queue for sorting by hits by terms in more than one field.
Creates a hit queue sorted by the given list of fields.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
A utility for making Lucene Documents from a File.
Returns true iff a file with the given name exists.
Returns true iff a file with the given name exists.
Returns true iff the named file exists in this directory.
A DocumentHandler implementation to delegate responsibility to
based on a files extension.
Returns the length of a file in the directory.
Returns the length in bytes of a file in the directory.
Returns the length in bytes of a file in the directory.
Returns the time the named file was last modified.
Returns the time the named file was last modified.
Returns the time the named file was last modified.
Returns the time the named file was last modified.
Abstract base class providing a mechanism to restrict searches to a subset
of an index.
A query that applies a filter to the results of another query.
Constructs a new query which applies a filter to the results of the original query.
Abstract class for enumerating a subset of all terms.
A FilterIndexReader
contains another IndexReader, which it
uses as its basic source of data, possibly transforming the data along the
way or providing additional functionality.
Construct a FilterIndexReader based on the specified base reader.
Base class for filtering
TermDocs
implementations.
Base class for filtering
TermEnum
implementations.
Release the write lock, if needed.
Release the write lock, if needed.
Utility class for executing code that needs to do
something with the current segments file.
Generated class implementing code defined by a snowball script.
Sort using term values as encoded Floats.
Interface to parse floats from document fields.
Converts a 32 bit float to an 8 bit float.
floatToByte(b, mantissaBits=3, zeroExponent=15)
smallest non-zero value = 5.820766E-10
largest value = 7.5161928E9
epsilon = 0.125
floatToByte(b, mantissaBits=5, zeroExponent=2)
smallest nonzero value = 0.033203125
largest value = 1984.0
epsilon = 0.03125
Forces any buffered output to be written.
Make sure all changes are written to disk.
Forces any buffered output to be written.
Flush all in-memory buffered updates (adds and deletes)
to the Directory.
Expert: implements buffer write.
Expert: implements buffer write.
Format - class org.apache.lucene.benchmark.byTask.utils.
Format Formatting utilities (for reports).
The file format version, a negative number.
Padd a string from right.
This format adds details used for lockless commits.
This is the current file format written.
Padd a number from right.
Processes terms found in the original text, typically by applying some form
of mark-up to highlight terms in HTML search results pages.
Implements the policy for breaking text into multiple fragments for consideration
by the
Highlighter
class.
Free memory at the end of measurement interval.
Extended list of typical French stopwords.
Analyzer for French language.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
A filter that stemms french words.
Generated class implementing code defined by a snowball script.
A stemmer for French words.
Returns the frequency of the term within the current document.
Returns the frequency of the term within the current document.
Get the n-gram from the front of the input
Straightforward implementation of
Directory
as a directory of files.
Delete files and directories, even if non-empty.
Delete files and directories, even if non-empty.
Implements the fuzzy search query.
Create a new FuzzyQuery that will match terms with a similarity
of at least minimumSimilarity
to term
.
Subclass of FilteredTermEnum for enumerating all terms that are similiar
to the specified filter term.
Creates a FuzzyTermEnum with an empty prefix and a minSimilarity of 0.5f.
Creates a FuzzyTermEnum with an empty prefix.
Constructor for enumeration of all terms from specified reader
which share a prefix of
length prefixLength
with term
and which have a fuzzy similarity >
minSimilarity
.
Generated class implementing code defined by a snowball script.
List of typical german stopwords.
Analyzer for German language.
Builds an analyzer with the default stop words
(GERMAN_STOP_WORDS
).
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
A filter that stems German words.
Builds a GermanStemFilter that uses an exclusiontable.
Generated class implementing code defined by a snowball script.
A stemmer for German words.
Return Lucene's package, including version information.
Returns true
if bit
is one and
false
if it is zero.
Returns the string value of the field with the given name if any exist in
this document, or null.
Returns the string value of the field with the given name if any exist in
this document, or null.
Return a boolean property.
Return a string property.
Prepare a list of benchmark data, using all possible combinations of
benchmark parameters.
Prepare a list of benchmark data, using all possible combinations of
benchmark parameters.
Returns the analyzer used by this index.
A convenience method that tries a number of approaches to getting a token stream.
Checks the internal cache for an appropriate entry, and if
none is found reads field
to see if it contains integers, floats
or strings, and then calls one of the other methods in this class to get the
values.
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the first character for current token (being
matched after the last call to BeginTOken).
Highlights chosen terms in a text, extracting the most relevant section.
Highlights chosen terms in a text, extracting the most relevant section.
Highlights chosen terms in a text, extracting the most relevant sections.
Highlights chosen terms in a text, extracting the most relevant sections.
Highlights chosen terms in a text, extracting the most relevant sections.
Highlights terms in the text , extracting the most relevant sections
and concatenating the chosen fragments with a separator (typically "...").
Low level api to get the most relevant (formatted) sections of the document.
Returns an array of bytes for the first (or only) field that has the name
specified as the method parameter.
Returns an array of byte arrays for of the fields that have the name specified
as the method parameter.
Gets the bodyText attribute of the
HtmlDocument
object.
Factory method for generating query, given a set of clauses.
Factory method for generating query, given a set of clauses.
Factory method for generating query, given a set of clauses.
Factory method for generating query, given a set of clauses.
Returns the boost factor for hits for this field.
Returns the boost factor for hits on any field of this document.
Returns the boost factor for hits for this field.
Returns the boost factor for this hit on any field of the underlying document.
Gets the boost for this clause.
Return total byte size of docs made since last reset.
Return total byte size of docs made since last reset.
Returns the set of clauses in this query.
Return the clauses whose spans are matched.
Return the clauses whose spans are matched.
Returns the column position of the character last read.
Returns the column position of the character last read.
Returns the column position of the character last read.
Returns the column position of the character last read.
Returns an object which, when sorted according to natural order,
will order the Term values in the correct order.
Return number of docs made since last reset.
Return number of docs made since last reset.
Get the segments_N filename in use by this segment infos.
Get the filename of the current segments_N file
in the directory.
Get the filename of the current segments_N file
from a list of files.
Get the generation (N) of the current segments_N file
in the directory.
Get the generation (N) of the current segments_N file
from a list of files.
Reads version number from segments files.
Reads version number from segments files.
Reads version number from segments files.
Checks the internal cache for an appropriate entry, and if none
is found reads the terms out of field
and calls the given SortComparator
to get the sort values.
Returns the date resolution that is used by RangeQueries for the given field.
Return the default Similarity implementation used by indexing and search
code.
Gets implicit operator setting, which will be either AND_OPERATOR
or OR_OPERATOR.
Gets implicit operator setting, which will be either AND_OPERATOR
or OR_OPERATOR.
A description of this explanation node.
The sub-nodes of this explanation node.
Returns the Directory used by this index.
Returns the directory instance for the named location.
Returns the directory instance for the named location.
Returns the directory instance for the named location.
Returns the directory instance for the named location.
Returns the directory instance for the named location.
Returns the directory instance for the named location.
Returns whether Lucene's use of lock files is disabled.
Returns document for this hit.
Gets the document attribute of the DocumentHandler object
Gets the document attribute of the
FileExtensionDocumentHandler object
Return the maximum end position permitted in a match.
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the column number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Returns the line number of the last character for current token (being
matched after the last call to BeginTOken).
Construct the enumeration to be used, expanding the pattern term.
Construct the enumeration to be used, expanding the pattern term.
Construct the enumeration to be used, expanding the pattern term.
Retrieves nested exception.
Return the SpanQuery whose matches must not overlap those returned.
Returns the field name for this query
Returns the field name for this query
Returns the name of the field.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns the name of the field matched by this query.
Returns a field with the given name if any exist in this document, or
null.
Returns a field with the given name if any exist in this document, or
null.
Returns an array of
Fieldable
s with the given name.
Get a list of unique field names that exist in this index and have the specified
field option information.
Returns a List of all the fields in a document.
Returns an array of
Field
s with the given name.
Returns the current position in this file, where the next read will
occur.
Returns the current position in this file, where the next write will
occur.
Returns the current position in this file, where the next read will
occur.
Returns the current position in this file, where the next write will
occur.
Returns the encapsulated filter
Checks the internal cache for an appropriate entry, and if
none is found, reads the terms in field
as floats and returns an array
of size reader.maxDoc()
of the value each document
has in the given field.
Checks the internal cache for an appropriate entry, and if
none is found, reads the terms in field
as floats and returns an array
of size reader.maxDoc()
of the value each document
has in the given field.
Called when the highlighter has no more tokens for the current fragment - the scorer returns
the weighting it has derived for the most recent fragment, typically based on the tokens
passed to getTokenScore().
Get the minimal similarity for fuzzy queries.
Get the minimal similarity for fuzzy queries.
Get the prefix length for fuzzy queries.
Get the prefix length for fuzzy queries.
Called when parser parses an input term token that has the fuzzy suffix (~) appended.
getId() - method in class org.apache.lucene.search.
Hit Returns id for this hit.
Extracts all terms texts of a given Query into an array of WeightedTerms
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Returns a string made up of characters from the marked token beginning
to the current buffer position.
Return the SpanQuery whose matches are filtered.
Checks the internal cache for an appropriate entry, and if none is
found, reads the terms in field
as integers and returns an array
of size reader.maxDoc()
of the value each document
has in the given field.
Checks the internal cache for an appropriate entry, and if none is found,
reads the terms in field
as integers and returns an array of
size reader.maxDoc()
of the value each document has in the
given field.
Short legend for interpreting toString() output.
Get a list of all available types of data points.
Get a short legend for toString() output.
Returns the line number of the character last read.
Returns the line number of the character last read.
Returns the line number of the character last read.
Returns the line number of the character last read.
Returns current locale, allowing access by subclasses.
Returns current locale, allowing access by subclasses.
Returns the Locale by which term values are interpreted.
Get the LockFactory that this Directory instance is
using for its locking implementation.
Return a string identifier that uniquely differentiates
this Directory instance from other Directory instances.
Return a string identifier that uniquely differentiates
this Directory instance from other Directory instances.
Get the prefix in use for all locks created in this LockFactory.
Returns a simple analyzer wrapper that logs all tokens produced by the
underlying child analyzer to the given log stream (typically System.err);
Otherwise behaves exactly like the child analyzer, delivering the very
same tokens; useful for debugging purposes on custom indexing and/or
querying.
How often to print out log messages when in benchmark loops
Returns the lower term of this range query
Returns the value of the lower endpoint of this range query, null if open ended
The match status of this explanation node.
Return the SpanQuery whose matches are filtered.
Return the maximum number of clauses permitted, 1024 by default.
returns the maximum score encountered by elements inserted via insert()
Expert: Returns the maximum score value encountered.
Returns an analyzer wrapper that returns at most the first
maxTokens
tokens from the underlying child analyzer,
ignoring all remaining tokens.
Returns a reasonable approximation of the main memory [bytes] consumed by
this instance.
This method has the standard behavior when this object has been
created using the standard constructors.
This method has the standard behavior when this object has been
created using the standard constructors.
This method has the standard behavior when this object has been
created using the standard constructors.
This method has the standard behavior when this object has been
created using the standard constructors.
This method has the standard behavior when this object has been
created using the standard constructors.
You can also modify the body of this method to customize your error messages.
You can also modify the body of this method to customize your error messages.
You can also modify the body of this method to customize your error messages.
You can also modify the body of this method to customize your error messages.
You can also modify the body of this method to customize your error messages.
Gets the minimum number of the optional BooleanClauses
which must be satisifed.
Returns the minimum similarity that is required for this query to match.
Returns (frequency:term) pairs for the top N distinct terms (aka words),
sorted descending by frequency (and ascending by term, if tied).
Return the data of the next document.
Return the data of the next document.
Return the data of the next document.
Return the data of the next document.
Get the next segments_N filename that will be written.
Returns a table for decoding normalization bytes.
Returns an array of TermVectorOffsetInfo in which the term is found.
Returns an array of TermVectorOffsetInfo in which the term is found.
True if norms are omitted for this indexed field
True if norms are omitted for this indexed field
Returns total number of stats points when this report was created.
Returns at most the first N paragraphs of the given text.
Gets the default slop for phrases.
Gets the default slop for phrases.
Returns an English stemming analyzer that stems tokens from the
underlying child analyzer according to the Porter stemming algorithm.
Returns the position increment of this Token.
Invoked before indexing a Fieldable instance if
terms have already been added to that field.
Return the positionIncrementGap from the analyzer assigned to fieldName
Returns the relative positions of terms in this phrase.
Returns the relative positions of terms in this phrase.
Returns the prefix of this query.
Returns the non-fuzzy prefix length.
Called when parser parses an input term
token that uses prefix notation; that is, contains a single '*' wildcard
character as its last character.
The query that this concerns.
Return query maker used for this task.
Return query maker used for this task.
Return query maker used for this task.
Overrides super class, by passing terms through analyzer.
Returns the rate per minute: how many operations should be performed in a minute.
Get rate of processing, defined as number of processed records per second.
Returns number of stats points represented in this report.
Returns whether the sort should be reversed.
The number of times to run the benchmark
Returns score for this hit.
Returns at most the first N sentences of the given text.
Expert: Return the Similarity implementation used by this IndexWriter.
Returns the Similarity implementation used by this scorer.
Expert: Return the Similarity implementation used by this Searcher.
Expert: Returns the Similarity implementation to be used for this query.
getSize() - method in class org.apache.lucene.benchmark.byTask.stats.
Report Returns number of lines in the reoprt.
Sets the phrase slop for this query.
Return the maximum number of intervening unmatched positions permitted.
Representation of the sort criteria.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Expert: Returns the matches for this query in an index.
Checks the internal cache for an appropriate entry, and if none
is found reads the term values in field
and returns
an array of them in natural order, along with an array telling
which element in the term array each document uses.
Checks the internal cache for an appropriate entry, and if none
is found, reads the term values in field
and returns an array
of size reader.maxDoc()
containing the value each document
has in the given field.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
Returns an array of characters that make up the suffix of length 'len' for
the currently matched token.
A short one line summary which should contain all high level
information about this Explanation, without the "Details"
Returns an analyzer wrapper that wraps the underlying child analyzer's
token stream into a
SynonymTokenFilter
.
Returns the synonym set for the given word, sorted ascending.
Returns the pattern term.
Return the term whose spans are matched.
Returns the term of this query.
Returns a List of the terms in the multiphrase.
Array of term frequencies.
Return a term frequency vector for the specified document and field.
Return an array of term frequency vectors for the specified document.
Return an array of term frequency vectors for the specified document.
Expert: Return the interval between indexed terms.
Returns an array of positions in which the term is found.
Returns an array of positions in which the term is found.
Returns the set of terms in this phrase.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Returns a collection of all terms matched by this query.
Extracts all terms texts of a given Query into an array of WeightedTerms
Extracts all terms texts of a given Query into an array of WeightedTerms
Extracts all terms texts of a given Query into an array of WeightedTerms
Similar to getAll(java.io.File[], org.apache.lucene.analysis.Analyzer[])
but only uses
maxBufferedDocs of 10 and 100 and same for mergeFactor, thus reducing the number of permutations significantly.
getText() - method in class org.apache.lucene.benchmark.byTask.stats.
Report Returns the report text.
Gets the title attribute of the HtmlDocument
object.
Returns an analyzer wrapper that caches all tokens generated by the underlying child analyzer's
token streams, and delivers those cached tokens on subsequent calls to
tokenStream(String fieldName, Reader reader)
if the fieldName has been seen before, altogether ignoring the Reader parameter on cache lookup.
Called for each token in the current fragment
The total number of documents that matched this query.
Get total values from all data points of all types.
Get total values from all data points of a given type.
Returns the type of contents in the field.
Returns the upper term of this range query
Returns the value of the upper endpoint of this range query, null if open ended
Get the current setting of whether to use the compound file format.
The value assigned to this explanation node.
The weight for this query.
Returns an array of values of the field specified as the method parameter.
Version number when this IndexReader was opened.
version number when this SegmentInfos was generated.
Called when parser
parses an input term token that contains one or more wildcard
characters (? and *), but is not a prefix term token (one
that has just a single * character at the end)
Depending on analyzer and settings, a wildcard term may (most probably will)
be lower-cased automatically.
Factory method for generating a query.
Factory method for generating a query.
Factory method for generating a query.
Loads a text file and adds every line as an entry to a HashSet (omitting
leading and trailing whitespace).
Reads lines from a Reader and adds every line as an entry to a HashSet (omitting
leading and trailing whitespace).
Return all words present in the dictionary
Return all words present in the dictionary
Return all words present in the dictionary
Formats text with different color intensity depending on the score of the
term.
Sets the color range for the IDF scores
Analyzer for the Greek language.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
GreekCharsets class contains encodings schemes (charsets) and toLowerCase() method implementation
for greek characters in Unicode, ISO-8859-7 and Microsoft Windows CP1253.
Normalizes token text to lower case, analyzing given ("greek") charset.
id - field in class org.apache.lucene.benchmark.stats.
QueryData Benchmark id
id(int) - method in class org.apache.lucene.search.
Hits Returns the id for the nth document in this set.
Computes a score factor for a phrase.
Implemented as log(numDocs/(docFreq+1)) + 1
.
Computes a score factor based on a term's document frequency (the number
of documents which contain the term).
Computes a score factor based on a term's document frequency (the number
of documents which contain the term).
Computes a score factor for a simple term.
image - field in class org.apache.lucene.analysis.standard.
Token The string image of the token.
image - field in class org.apache.lucene.demo.html.
Token The string image of the token.
image - field in class org.apache.lucene.queryParser.
Token The string image of the token.
image - field in class org.apache.lucene.queryParser.precedence.
Token The string image of the token.
image - field in class org.apache.lucene.queryParser.surround.parser.
Token The string image of the token.
Returns true
if the lower endpoint is inclusive
Returns true
if the upper endpoint is inclusive
Specifies whether and how a field should be indexed.
Just like indexOf(int)
but searches for a number of terms
at the same time.
Returns true
if an index exists at the specified directory.
Returns true
if an index exists at the specified directory.
Returns true
if an index exists at the specified directory.
Filename filter that accept filenames and extensions only created by Lucene.
Index all text files under a directory.
Abstract base class for input from a file in a
Directory
.
Merges indices specified on the command line into the index
specified as the first command line argument.
[Note that as of
2.1, all but one of the
methods in this class are available via
IndexWriter
.
Open an index with write access.
Open an index with write access.
Open an index with write access.
Return an index in the term numbers array returned from
getTerms
at which the term with the specified
term
appears.
Special comparator for sorting hits according to index order (document number).
Represents sorting by index order.
Abstract base class for output to a file in a Directory.
IndexReader is an abstract class, providing an interface for accessing an
index.
Constructor used if IndexReader is not owner of its directory.
Implements search over a single IndexReader.
Creates a searcher searching the provided index.
Creates a searcher searching the index in the provided directory.
Creates a searcher searching the index in the named directory.
Ant task to index files with Lucene
An IndexWriter creates and maintains an index.
Constructs an IndexWriter for the index in
path
, creating it first if it does not
already exist, otherwise appending to the existing
index.
Constructs an IndexWriter for the index in path
.
Constructs an IndexWriter for the index in
d
, creating it first if it does not
already exist, otherwise appending to the existing
index.
Constructs an IndexWriter for the index in d
.
Constructs an IndexWriter for the index in
path
, creating it first if it does not
already exist, otherwise appending to the existing
index.
Constructs an IndexWriter for the index in path
.
Initialize an IndexWriter.
Subclass constructors must call this.
The source of tokens for this filter.
The text source for this Tokenizer.
Adds element to the PriorityQueue in log(size) time if either
the PriorityQueue is not full, or not lessThan(element, top()).
Adds a Scorer to the ScorerDocQueue in log(size) time if either
the ScorerDocQueue is not full, or not lessThan(scorer, top()).
INT - static field in class org.apache.lucene.search.
SortField Sort using term values as encoded Integers.
Interface to parse ints from document fields.
True iff the value of the filed is stored as binary
True iff the value of the filed is stored as binary
Returns true if this is a file that would be contained
in a CFS file.
True if the value of the field is stored and compressed within the index
True if the value of the field is stored and compressed within the index
Check whether this IndexReader still works on a current version of the index.
Returns true if document n has been deleted
Returns true
if the range query is inclusive
True iff the value of the field is to be indexed, so that it may be
searched on.
True iff the value of the field is to be indexed, so that it may be
searched on.
Return true if matches are required to be in-order.
Indicates whether a Field is Lazy or not.
Returns true if the resource is currently locked.
Returns true
iff the index in the named directory is
currently locked.
Returns true
iff the index in the named directory is
currently locked.
Indicates wether or not this Explanation models a good match.
Indicates wether or not this Explanation models a good match.
Test to see if this token from the stream should be held in a new TextFragment
Test to see if this token from the stream should be held in a new TextFragment
A filter that replaces accented characters in the ISO Latin 1 character set
(ISO-8859-1) by their unaccented equivalent.
Checks is the index is optimized (if it has a single segment and no deletions)
True iff the value of the field is to be stored in the index for return
with search hits.
True iff the value of the field is to be stored in the index for return
with search hits.
True iff terms are stored as term vector together with their offsets
(start and end positon in source text).
True iff terms are stored as term vector together with their offsets
(start and end positon in source text).
True iff terms are stored as term vector together with their token positions.
True iff terms are stored as term vector together with their token positions.
Returns true iff a character should be included in a token.
Collects only characters which satisfy
Character.isLetter(char)
.
Collects only characters which satisfy
Character.isLetter(char)
.
Collects only characters which do not satisfy
Character.isWhitespace(char)
.
True iff the value of the field should be tokenized as text prior to
indexing.
True iff the value of the field should be tokenized as text prior to
indexing.
Generated class implementing code defined by a snowball script.
An Iterator over the disjuncts
name - field in class org.apache.lucene.benchmark.stats.
TimeData Name of the data point - usually one of a data series with the same name
Returns the name of the field as an interned string.
Returns the name of the field as an interned string.
Implements
LockFactory
using native OS file locks
(available through java.nio.*).
Create a NativeFSLockFactory instance, storing lock
files into the specified lockDir:
Create a NativeFSLockFactory instance, storing lock
files into the specified lockDirName:
Snowball generated stemmer classes.
NEW_LINE - static field in class org.apache.lucene.benchmark.byTask.tasks.
PerfTask Creates a comparator for the field in the given index.
Increment the round number, for config values that are extracted by round number.
Increment the counter for properties maintained by Round Number.
Returns a new Token object, by default.
Returns a new Token object, by default.
Returns a new Token object, by default.
Returns a new Token object, by default.
newToken(int) - static method in class org.apache.lucene.queryParser.surround.parser.
Token Returns a new Token object, by default.
next - field in class org.apache.lucene.analysis.standard.
Token A reference to the next regular (non-special) token from the input
stream.
next - field in class org.apache.lucene.demo.html.
Token A reference to the next regular (non-special) token from the input
stream.
next - field in class org.apache.lucene.queryParser.
Token A reference to the next regular (non-special) token from the input
stream.
next - field in class org.apache.lucene.queryParser.precedence.
Token A reference to the next regular (non-special) token from the input
stream.
next - field in class org.apache.lucene.queryParser.surround.parser.
Token A reference to the next regular (non-special) token from the input
stream.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Increments the enumeration to the next element.
Moves to the next pair in the enumeration.
Increments the enumeration to the next element.
Returns the next token in the stream, or null at EOS.
Returns a
Hit
instance representing the next hit in
Hits
.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next input Token whose termText() is the right len
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next input Token, after being stemmed
Advances to the next document matching the query.
Advances to the next document matching the query.
Returns the next input Token, after being stemmed
next() - method in class org.apache.lucene.search.spans.
Spans Move to the next match, returning true iff any such exists.
Returns the next token in the stream, or null at EOS.
Returns the next token in the stream, or null at EOS.
Returns the next input Token whose termText() is not a stop word.
Returns the next token in the stream, or null at EOS.
Moves to the next pair in the enumeration.
Increments the enumeration to the next element.
Returns the next token in the stream, or null at EOS.
Returns next position in the current document.
Returns next position in the current document.
Tokenizes the input into n-grams of the given size(s).
Creates NGramTokenizer with default min and max n-grams.
Creates NGramTokenizer with given min and max n-grams.
NO - static field in class org.apache.lucene.document.
Field.Index Do not index the field value.
NO - static field in class org.apache.lucene.document.
Field.Store Do not store the field value in the index.
Do not store term vectors.
Index the field's value without an Analyzer, and disable
the storing of norms.
"\\W+"
; Divides text at non-letters (!Character.isLetter(c))
Called on each token character to normalize it before it is added to the
token.
Collects only characters which satisfy
Character.isLetter(char)
.
Assigns the query normalization factor to this.
Returns the byte-encoded normalization factor for the named field of
every document.
Reads the byte-encoded normalization factor for the named field of every
document.
Generated class implementing code defined by a snowball script.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Fragmenter
implementation which does not fragment the text.
Provides support for converting longs to Strings, and back again.
Returns the number of documents in this index.
Expert: Return the number of documents whose segments are currently cached in memory.
Return total bytes of all available unique texts, 0 if not applicable
Return total bytes of all available unique texts, 0 if not applicable
Return how many real unique texts are available, 0 if not applicable.
Return how many real unique texts are available, 0 if not applicable.
Return how many real unique texts are available, 0 if not applicable.
Return how many real unique texts are available, 0 if not applicable.
Creates a new
RAMDirectory
instance from the
FSDirectory
.
Creates a new RAMDirectory
instance from a different
Directory
implementation.
Creates a new
RAMDirectory
instance from the
FSDirectory
.
Construct an empty output buffer.
Expert: Return the total size of all index files currently cached in memory.
A Filter that restricts search results to a range of values in a given
field.
A Query that matches documents within an exclusive range.
Constructs a query selecting all terms greater than
lowerTerm
but less than upperTerm
.
Attempts to read multiple entries from the enumeration, up to length of
docs.
Attempts to read multiple entries from the enumeration, up to length of
docs.
This version of read uses the retry logic (for lock-less
commits) to find the right segments file to load.
Read a particular segmentFileName.
Reads and returns a single byte.
Reads and returns a single byte.
Reads a specified number of bytes into an array at the specified offset.
Reads a specified number of bytes into an array at the specified offset.
Returns the next character from the selected input.
Returns the next character from the selected input.
Returns the next character from the selected input.
Returns the next character from the selected input.
Returns the next character from the selected input.
Returns the next character from the selected input.
Returns the next character from the selected input.
Reads UTF-8 encoded characters into an array.
Current version number from segments file.
The value of the field as a Reader, or null.
The value of the field as a Reader, or null.
Reads four bytes and returns an int.
Expert: implements buffer refill.
Reads eight bytes and returns a long.
Resolves the deserialized instance to the local reference for accurate
equals() and == comparisons.
Read index (abstract) task.
Reads an int stored in variable-length format.
Reads a long stored in variable-length format.
This class exists as a gateway to access useful Jakarta Regexp package protected data.
Implements the regular expression term search query.
Constructs a query for terms matching term
.
Subclass of FilteredTermEnum for enumerating all terms that match the
specified regular expression term using the specified regular expression
implementation.
Releases exclusive access.
Special comparator for sorting hits according to computed relevance (document score).
Represents sorting by computed relevance.
A remote searchable implementation.
Constructs and exports a remote searcher.
To replace accented characters in a String by unaccented equivalents.
Removes field with the specified name from the document.
Removes all fields with the given name from the document.
Renames an existing file in the directory.
Renames an existing file in the directory.
Renames an existing file in the directory.
If true, re-open index reader before benchmark.
Report all statistics with no aggregations.
Report - class org.apache.lucene.benchmark.byTask.stats.
Report Textual report of current statistics.
Report detailed statistics as a string
Report statistics as a string, aggregate for tasks named the same.
Report statistics as a string, aggregate for tasks named the same, and from the same round.
Report (abstract) task - all report tasks extend this task.
Report by-name-prefix statistics with no aggregations.
Report all statistics grouped/aggregated by name and round.
Report all statistics aggregated by name.
Report all prefix matching statistics grouped/aggregated by name and round.
Report by-name-prefix statistics aggregated by name.
A Scorer for queries with a required subscorer and an excluding (prohibited) subscorer.
Construct a ReqExclScorer
.
A Scorer for queries with a required part and an optional part.
Construct a ReqOptScorer
.
Resets this to an empty buffer.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise, as if it just started.
Reset inputs so that the test run would behave, input wise,
as if it just started.
Resets the norms for the specified field.
Resets the norms for the specified field.
Reset all index and input data and call gc, erase index and dir, does NOT clear statistics.
Reset all index and input data and call gc, does NOT erase index/dir, does NOT clear statistics.
Specifies the time granularity.
If true, actually retrieve documents returned in Hits.
A DocMaker using the Reuters collection for its input.
A QueryMaker that makes queries devised manually (by Grant Ingersoll) for
searching in the Reuters collection.
Expert: called to re-write queries into primitive queries.
Optimize our representation and our subqueries representations
Rewrites the wrapped query.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Expert: called to re-write queries into primitive queries.
Limit a date's resolution.
Limit a date's resolution.
Run the task, record statistics.
Run benchmark using supplied parameters.
Analyzer for Russian language.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
RussianCharsets class contains encodings schemes (charsets) and toLowerCase() method implementation
for russian characters in Unicode, KOI8 and CP1252.
A RussianLetterTokenizer is a tokenizer that extends LetterTokenizer by additionally looking up letters
in a given "russian charset".
Normalizes token text to lower case, analyzing given ("russian") charset.
A filter that stems Russian words.
Generated class implementing code defined by a snowball script.
s - field in class net.sf.snowball.
Among Sample - class org.apache.lucene.benchmark.byTask.programmatic.
Sample Sample performance test written programatically - no algorithm file is needed here.
Expert: The score of this document for the query.
Sort by document score (relevancy).
Returns the score of the current document matching the query.
Returns the score of the current document matching the query.
Returns the score of the current document matching the query.
Returns the score for the nth document in this set.
Scores and collects all matching documents.
Expert: Collects matching documents in a range.
Expert: Returned by low-level search implementations.
Expert: Constructs a ScoreDoc.
Expert: Compares two ScoreDoc objects for sorting.
Expert: The top hits for the query.
Expert: Common scoring functionality for different types of queries.
Scorer - interface org.apache.lucene.search.highlight.
Scorer Adds to the score for a fragment based on its tokens
Constructs a scorer for this.
A ScorerDocQueue maintains a partial ordering of its Scorers such that the
least Scorer can always be found in constant time.
Create a ScorerDocQueue with a maximum size.
Convenience method that efficiently returns the relevance score by
matching this index against the given Lucene query expression.
Returns the documents matching query
.
Returns the documents matching query
and
filter
.
Expert: Low-level search implementation.
Expert: Low-level search implementation with arbitrary sorting.
Returns documents matching query
and filter
,
sorted by sort
.
Returns documents matching query
sorted by
sort
.
Expert: Low-level search implementation.
A search implementation which spans a new thread for each
Searchable, waits for each search to complete and merge
the results back together.
Expert: Low-level search implementation.
Expert: Low-level search implementation.
Expert: Low-level search implementation with arbitrary sorting.
A search implementation allowing sorting which spans a new thread for each
Searchable, waits for each search to complete and merges
the results back together.
Expert: Low-level search implementation with arbitrary sorting.
Expert: Low-level search implementation with arbitrary sorting.
The interface for search implementations.
An abstract base class for search implementations.
Simple command-line based search demo.
Search and Travrese and Retrieve docs task.
Search and Travrese task.
Sets current position in this file, where the next read will occur.
Sets current position in this file, where the next write will occur.
Sets current position in this file, where the next read will occur.
Sets current position in this file, where the next write will occur.
Sets current position in this file, where the next write will occur.
Sets this to the data for a term.
Sets this to the data for a term.
Sets this to the data for the current term in a
TermEnum
.
Sets this to the data for the current term in a
TermEnum
.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Sets the value of bit
to one.
Sets the accuracy 0 < minScore < 1; default 0.5
Set to true
to allow *
and ?
as the first character
of a PrefixQuery and WildcardQuery.
Sets the analyzer based on the builtin Lucene analyzer types.
Declare what fields to load normally and what fields to load lazily
Pass in the Set of
Field
names to load and the Set of
Field
names to load lazily.
Sets the baseline and minimum function variables for baselineTf
Sets the boost factor hits on this field.
Sets a boost factor for hits on any field of this document.
Sets the boost factor hits on this field.
Sets the boost for this query clause to b
.
Sets the default date resolution used by RangeQueries for fields for which no
specific date resolutions has been set.
Sets the date resolution used by RangeQueries for a specific field.
Set the default Similarity implementation used by indexing and search
code.
Advanced: set how many times to try loading the
segments.gen file contents to determine current segment
generation.
Advanced: set how many milliseconds to pause in between
attempts to load the segments.gen file.
Advanced: set how many times to try incrementing the
gen when loading the segments file.
Sets the boolean operator of the QueryParser.
Sets the boolean operator of the QueryParser.
Sets the default (for any instance of IndexWriter) maximum time to wait for a write lock (in
milliseconds).
Sets the description of this explanation node.
Set whether Lucene's use of lock files is disabled.
Sets the documentHandler attribute of the IndexTask object
Set an alternative exclusion list for this filter.
Set an alternative exclusion list for this filter.
Set an alternative exclusion list for this filter.
Set the minimum similarity for fuzzy queries.
Set the minimum similarity for fuzzy queries.
Set the prefix length for fuzzy queries.
Set the prefix length for fuzzy queries.
Sets the function variables for the hyperbolicTf functions
Specifies the directory where the index will be stored
If non-null, information about merges and a message when
getMaxFieldLength()
is reached will be printed to this.
If non-null, information about merges and a message when
maxFieldLength is reached will be printed to this.
If non-null, information about retries when loading
the segments file will be printed to this.
Sets the default function variables used by lengthNorm when no field
specifc variables have been set.
Sets the function variables used by lengthNorm for a specific named field
Set locale used by date range parsing.
Set locale used by date range parsing.
Set the LockFactory that this Directory instance should
use for its locking implementation.
Set the prefix in use for all locks created in this
LockFactory.
Whether terms of wildcard, prefix, fuzzy and range queries are to be automatically
lower-cased or not.
Whether terms of wildcard, prefix, fuzzy and range queries are to be automatically
lower-cased or not.
Sets the match status assigned to this explanation node.
Determines the minimal number of delete terms required before the buffered
in-memory delete terms are applied and flushed.
Determines the minimal number of documents required before the buffered
in-memory documents are merging and a new Segment is created.
Determines the minimal number of documents required before the buffered
in-memory documents are merged and a new Segment is created.
Set the maximum number of clauses permitted per BooleanQuery.
The maximum number of terms that will be indexed for a single field in a
document.
The maximum number of terms that will be indexed for a single field in a
document.
Determines the largest number of documents ever merged by addDocument().
Expert: Sets the maximum score value encountered.
Determines how often segment indices are merged by addDocument().
Sets the mergeFactor attribute of the IndexTask object
Determines how often segment indices are merged by addDocument().
Specifies a minimum number of the optional BooleanClauses
which must be satisifed.
Execute child tasks in a way that they do not reprt their time separately.
Expert: Resets the normalization factor for the named field of the named
document.
Expert: Resets the normalization factor for the named field of the named
document.
Expert:
If set, omit normalization factors associated with this indexed field.
Expert:
If set, omit normalization factors associated with this indexed field.
Sets the overwrite attribute of the IndexTask object
Set the params (docSize only)
Set the params (docid only)
Set the params of this task.
Set the params (property name and value).
Sets the default slop for phrases.
Sets the default slop for phrases.
Set the position increment.
Set a performance test configuration property.
Defines which RegexCapabilities
implementation is used by this instance.
Expert: Set the Similarity implementation used by this IndexWriter.
Expert: Set the Similarity implementation used by this Searcher.
Sets the phrase slop for this query.
Sets the number of other words permitted between words in query phrase.
Sets the sort to the given criteria.
Sets the sort to the given criteria in succession.
Sets the sort to the terms in field
then by index order
(document number).
Sets the sort to the terms in field
possibly in reverse,
then by index order (document number).
Sets the sort to the terms in each field in succession.
Reads a stemdictionary file , that overrules the stemming algorithm
This is a textfile that contains per line
word\tstem, i.e: two tab seperated words
Set dictionary for stemming, this dictionary overrules the algorithm,
so you can correct for a particular unwanted word-stem pair.
Builds an exclusionlist from the words contained in the given file.
Builds an exclusionlist from the words contained in the given file.
Builds an exclusionlist from the words contained in the given file.
Builds an exclusionlist from the words contained in the given file.
Builds an exclusionlist from a Hashtable.
Builds an exclusionlist from a Hashtable.
Builds an exclusionlist from a Hashtable.
Builds an exclusionlist from a Hashtable.
Builds an exclusionlist from an array of Strings.
Builds an exclusionlist from an array of Strings.
Builds an exclusionlist from an array of Strings.
Builds an exclusionlist from an array of Strings.
Set a alternative/custom GermanStemmer for this filter.
Set a alternative/custom FrenchStemmer for this filter.
Set a alternative/custom DutchStemmer for this filter.
Set a alternative/custom RussianStemmer for this filter.
Set the table model used by this table model
Expert: Set the interval between indexed terms.
Sets the Token's term text.
Task setup work that should not be measured for that specific task.
setup() - method in class org.apache.lucene.benchmark.byTask.tasks.
PerfTask Task setup work that should not be measured for that specific task.
Setting to turn on usage of a compound file.
Setting to turn on usage of a compound file.
If creating a new index and this is set to true, the
index will be created in compound format.
By default QueryParser uses new ConstantScoreRangeQuery in preference to RangeQuery
for range queries.
Sets the value assigned to this explanation node.
Sets the maximum time to wait for a write lock (in milliseconds) for this instance of IndexWriter.
Use this operator for clauses that should appear in the
matching documents.
Tasks that should never log at start can overide this.
Tasks that should never log at start can overide this.
Tasks that should not record statistics can overide this.
Tasks that should not record statistics can overide this.
Tasks that should not record statistics can overide this.
Tasks that should not record statistics can overide this.
Specifies which side of the input the n-gram should be generated from
Expert: Delegating scoring implementation.
Construct a
Similarity
that delegates all methods to another.
An Analyzer that filters LetterTokenizer with LowerCaseFilter.
An implementation of interface CharStream, where the stream is assumed to
contain only ASCII characters (without unicode processing).
Create documents for the test
Fragmenter
implementation which breaks text up into same-size
fragments with no concerns over spotting sentence boundaries.
Instantiate using the provided directory (as a File instance).
Instantiate using the provided directory name (String).
Simple
Encoder
implementation to escape text for HTML output
Simple
Formatter
implementation to highlight terms with a pre and post tag
Default constructor uses HTML: <B> tags to markup terms
Extract simple class name
A QueryMaker that makes queries for a collection created
using
SimpleDocMaker
.
Create sloppy phrase queries for performance test, in an index created using simple doc maker.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Implements
LockFactory
for a single in-process instance,
meaning all locking will take place through this one instance.
Expert: Load the size of this
Field
rather than its value.
Returns the number of bits in this vector.
Returns the number of elements currently stored in the PriorityQueue.
Returns the number of scorers currently stored in the ScorerDocQueue.
Expert: Like
SIZE
but immediately break from the field loading loop, i.e.
Return total size in bytes of all files in this
directory.
Skips entries to the first beyond the current whose document number is
greater than or equal to target.
Skips to the first match beyond the current whose document number is
greater than or equal to a given target.
Skips to the first match beyond the current whose document number is
greater than or equal to a given target.
Skips to the first match beyond the current whose document number is
greater than or equal to a given target.
Skips to the first match beyond the current, whose document number is
greater than or equal to target.
Skips entries to the first beyond the current whose document number is
greater than or equal to target.
Skips terms to the first beyond the current whose value is
greater or equal to target.
Implemented as 1 / (distance + 1)
.
Computes the amount of a sloppy phrase match, based on an edit distance.
Computes the amount of a sloppy phrase match, based on an edit distance.
Floating point numbers smaller than 32 bits.
Builds the named analyzer with no stop words.
Builds the named analyzer with the given stop words.
A filter that stems words using a Snowball-generated stemmer.
Construct the named stemming filter.
Sort - class org.apache.lucene.search.
Sort Encapsulates sort criteria for returned hits.
Sort() - constructor for class org.apache.lucene.search.
Sort Sorts by computed relevance.
Sorts by the criteria in the given SortField.
Sorts in succession by the criteria in each SortField.
Sorts by the terms in field
then by index order (document
number).
Sorts possibly in reverse by the terms in field
then by
index order (document number).
Sorts in succession by the terms in each field.
Abstract base class for sorting hits returned by a Query.
Expert: returns a comparator for sorting ScoreDocs.
Stores information about how to sort documents by terms in an individual
field.
Creates a sort by terms in the given field where the type of term value
is determined dynamically (
AUTO
).
Creates a sort, possibly in reverse, by terms in the given field where
the type of term value is determined dynamically (
AUTO
).
Creates a sort by terms in the given field with the type of term
values explicitly given.
Creates a sort, possibly in reverse, by terms in the given field with the
type of term values explicitly given.
Creates a sort by terms in the given field sorted
according to the given locale.
Creates a sort, possibly in reverse, by terms in the given field sorted
according to the given locale.
Creates a sort with a custom comparison function.
Creates a sort, possibly in reverse, with a custom comparison function.
Returns the type of sort.
Returns the value used to sort the given document.
Matches spans near the beginning of a field.
Construct a SpanFirstQuery matching spans in match
whose end
position is less than or equal to end
.
Formats text with different color intensity depending on the score of the
term using the span tag.
Generated class implementing code defined by a snowball script.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Matches spans which are near one another.
Construct a SpanNearQuery.
Removes matches which overlap with another SpanQuery.
Construct a SpanNotQuery matching spans from include
which
have no overlap with spans from exclude
.
Matches the union of its clauses.
Construct a SpanOrQuery merging the provided clauses.
Base class for span-based queries.
A SpanQuery version of
RegexQuery
allowing regular expression
queries to be nested within other SpanQuery subclasses.
Spans - interface org.apache.lucene.search.spans.
Spans Expert: an enumeration of span matches.
Matches spans containing a term.
Construct a SpanTermQuery matching the named term's spans.
This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
Spell Checker class (Main class)
(initially inspired by the David Spencer code).
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Builds an analyzer with the default stop words (
STOP_WORDS
).
Builds an analyzer with the stop words from the given file.
Builds an analyzer with the stop words from the given reader.
Builds an analyzer with the given stop words.
Builds an analyzer with the given stop words.
Reads in the Reuters Collection, downloaded from http://www.daviddlewis.com/resources/testcollections/reuters21578/reuters21578.tar.gz
in the workingDir/reuters and indexes them using the
StandardAnalyzer
Runs a standard set of documents through an Indexer and then runs a standard set of queries against the index.
A grammar-based tokenizer constructed with JavaCC.
Constructs a tokenizer for this Reader.
start() - method in class org.apache.lucene.search.spans.
Spans Returns the start position of the current match.
Start counting elapsed time.
Initializes the Fragmenter
Initializes the Fragmenter
called when a new fragment is started for consideration
Returns this Token's starting offset, the position of the first character
corresponding to this token in the source text.
Mark the starting time of this test run.
Stemms the given term to an unique discriminator.
Stemms the given term to a unique discriminator.
Stemms the given term to an unique discriminator.
Stop counting elapsed time.
An array containing some common English words that are not usually
useful for searching and some double-byte interpunctions.
An array containing some common English words that are usually not
useful for searching.
Filters LetterTokenizer with LowerCaseFilter and StopFilter.
Builds an analyzer which removes words in ENGLISH_STOP_WORDS.
Builds an analyzer with the stop words from the given file.
Builds an analyzer with the stop words from the given reader.
Builds an analyzer with the stop words from the given set.
Builds an analyzer which removes words in the provided array.
Removes stop words from a token stream.
Constructs a filter which removes words from the input
TokenStream that are named in the Set.
Construct a token stream filtering the given input.
Construct a token stream filtering the given input.
Constructs a filter which removes words from the input
TokenStream that are named in the array of words.
Specifies whether and how a field should be stored.
Sort using term values as Strings.
Indicator for StringIndex values in the cache.
Compares two strings, character by character, and returns the
first position where the two strings differ from one another.
Methods for manipulating strings.
Expert: Stores term text values and document ordering data.
Creates one of these objects
Converts a string-encoded date into a Date object.
Converts a string produced by timeToString
or
dateToString
back to a time, represented as a
Date object.
Converts a string-encoded date into a millisecond time.
Converts a string produced by timeToString
or
dateToString
back to a time, represented as the
number of milliseconds since January 1, 1970, 00:00:00 GMT.
The value of the field as a String, or null.
The value of the field as a String, or null.
Returns the document number of document n
within its
sub-index.
Returns index of the searcher for document n
in the array
used to construct this searcher.
Suggest similar words (restricted or not to a field of a user index)
The sum of squared weights of contained query clauses.
True iff running on SunOS.
Generated class implementing code defined by a snowball script.
A similarity with a lengthNorm that provides for a "platuea" of
equally good lengths, and tf helper functions.
Expand a query by looking up synonyms for every term.
Test program to look up synonyms.
The Token.type used to indicate a synonym to higher level filters.
Loads the
WordNet prolog file
wn_s.pl
into a thread-safe main-memory hash map that can be used for fast
high-frequency lookups of synonyms for any given (lowercase) word string.
Constructs an instance, loading WordNet synonym data from the given input
stream.
Injects additional tokens for synonyms of token terms fetched from the
underlying child stream; the child stream must deliver lowercase tokens
for synonyms to be found.
Creates an instance for the given underlying stream and synonym table.
The inner table model we are decorating
This is a TableModel that encapsulates Lucene
search logic within a TableModel implementation.
Tags - class org.apache.lucene.demo.html.
Tags Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Sequence of parallel or sequential tasks.
Statistics for a task run.
Task tearDown work that should not be measured for that specific task.
Task tearDown work that should not be measured for that specific task.
Term - class org.apache.lucene.index.
Term A Term represents a word from text.
Returns the current Term in the enumeration.
Returns the current Term in the enumeration.
Returns the current Term in the enumeration.
Constructs a Term with the given field and text.
Equality compare on the term
The termCompare method in FuzzyTermEnum uses Levenshtein distance to
calculate the distance between the given term and the comparing term.
Equality compare on the term
Equality compare on the term
TermDocs provides an interface for enumerating <document, frequency>
pairs for a term.
Returns an unpositioned
TermDocs
enumerator.
Returns an enumeration of all the documents which contain
term
.
Abstract class for enumerating terms.
Provides access to stored term vector of
a document field.
TermPositions provides an interface for enumerating the <document,
frequency, <position>* > tuples for a term.
Returns an enumeration of all the documents which contain
term
.
Extends TermFreqVector
to provide additional information about
positions in which each of the terms is found.
A Query that matches documents containing a term.
Constructs a query for the term t
.
Returns an enumeration of all the terms in the index.
Returns an enumeration of all terms after a given term.
Returns the Token's term text.
Specifies whether and how a field should have term vectors.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
This class holds together all parameters related to a test.
This class holds series of TimeData related to a single test run.
text() - method in class org.apache.lucene.index.
Term Returns the text of this term.
A utility for making Lucene Documents from a File.
Constructor for the TextDocument object
Low-level class used to record information about a section of a document
with a score.
Implemented as sqrt(freq)
.
Computes a score factor based on a term or phrase's frequency in a
document.
Computes a score factor based on a term or phrase's frequency in a
document.
Computes a score factor based on a term or phrase's frequency in a
document.
Analyzer for Thai language.
TokenFilter that use java.text.BreakIterator to break each
Token that is Thai into separate Token(s) for each Thai word.
This class holds a data point measuring speed of processing.
Converts a millisecond time to a string suitable for indexing.
Converts a millisecond time to a string suitable for indexing.
Render an explanation as HTML.
A Token is an occurence of a term from the text of a field.
Token - class org.apache.lucene.analysis.standard.
Token Describes the input token stream.
Describes the input token stream.
Describes the input token stream.
Token - class org.apache.lucene.queryParser.precedence.
Token Describes the input token stream.
Token - class org.apache.lucene.queryParser.surround.parser.
Token Describes the input token stream.
token - field in class org.apache.lucene.queryParser.surround.parser.
QueryParser Constructs a Token with the given term text, and start & end offsets.
Constructs a Token with the given text, start and end offsets, & type.
A TokenFilter is a TokenStream whose input is another token stream.
Construct a token stream filtering the given input.
One, or several overlapping tokens, along with the score(s) and the
scope of the original text
This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred.
Index the field's value so it can be searched.
A Tokenizer is a TokenStream whose input is a Reader.
Construct a tokenizer with null input.
Construct a token stream processing the given input.
Hides implementation issues associated with obtaining a TokenStream for use with
the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
from Analyzer class reparsing the stored content.
A TokenStream enumerates the sequence of tokens, either from
fields of a document or from query text.
Creates a TokenStream which tokenizes all the text in the provided
Reader.
Creates a TokenStream which tokenizes all the text in the provided Reader.
Creates a TokenStream which tokenizes all the text in the provided Reader.
get token stream from input
Creates a TokenStream which tokenizes all the text in the provided Reader.
Creates a TokenStream which tokenizes all the text in the provided TextReader.
Creates a TokenStream which tokenizes all the text in the provided Reader.
Creates a TokenStream which tokenizes all the text in the provided Reader.
Creates a TokenStream which tokenizes all the text in the provided Reader.
Creates a TokenStream which tokenizes all the text in the provided
Reader.
Creates a token stream that tokenizes all the text in the given Reader;
This implementation forwards to tokenStream(String, String)
and is
less efficient than tokenStream(String, String)
.
Creates a TokenStream which tokenizes all the text in the provided Reader.
Filters LowerCaseTokenizer with StopFilter.
Creates a TokenStream which tokenizes all the text in the provided
Reader.
Creates a token stream that tokenizes the given string into token terms
(aka words).
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Thrown when an attempt is made to add more than getMaxClauseCount()
clauses.
Returns the least element of the PriorityQueue in constant time.
Returns the least Scorer of the ScorerDocQueue in constant time.
Returns document number of the least Scorer of the ScorerDocQueue
in constant time.
A
HitCollector
implementation that collects the top-scoring
documents, returning them as a
TopDocs
.
Construct to collect a given number of hits.
Expert: Returned by low-level search implementations.
Construct to collect a given number of hits.
Expert: Returned by low-level sorted search implementations.
Scale down the values by divisor, append the unit string.
Prints a Field for human consumption.
Prints the fields of a document for human consumption.
Render an explanation as text.
Prints the parameters to be used to discover the promised result.
Returns a String representation of the index data for debugging purposes.
Prints a user-readable version of this query.
Prints a query to a string.
Returns a String representation of the index data for debugging purposes.
Return a string representation.
toString() - method in class org.apache.lucene.queryParser.precedence.
Token Returns the image.
toString() - method in class org.apache.lucene.queryParser.surround.parser.
Token Returns the image.
Return a tab-seprated string containing this data.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a user-readable version of this query.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a user-readable version of this query.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a query to a string, with field
assumed to be the
default field and omitted.
Prints a user-readable version of this query.
Licensed to the Apache Software Foundation (ASF) under one or more
contributor license agreements.
Expert: The total number of hits for the query.
Total memory at the end of measurement interval.
Set the modified time of an existing file to now.
Set the modified time of an existing file to now.
Set the modified time of an existing file to now.
A DocMaker using the (compressed) Trec collection for its input.
type() - method in class org.apache.lucene.analysis.
Token Returns this Token's lexical type.