Documentation
¶
Index ¶
- func Count(doc *documents.Document, tokeniser Tokeniser) *indices.InfoAndTerms
- func CountInDocuments(docs <-chan *documents.Document, tokeniser Tokeniser, ...)
- type EnglishTokeniser
- func (e *EnglishTokeniser) GetTerms(text string, operation func(string))
- func (e *EnglishTokeniser) IsStopWord(word string) bool
- func (e *EnglishTokeniser) Normalise(token string) string
- func (e *EnglishTokeniser) NormaliseMany(tokens []string) []string
- func (e *EnglishTokeniser) Tokenise(text string) []string
- type Tokeniser
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func CountInDocuments ¶
Types ¶
type EnglishTokeniser ¶
type EnglishTokeniser struct {
// contains filtered or unexported fields
}
func NewEnglishTokeniser ¶
func NewEnglishTokeniser(stopWordList io.Reader) (*EnglishTokeniser, error)
func NewEnglishTokeniserFromFile ¶
func NewEnglishTokeniserFromFile(stopWordFile string) (*EnglishTokeniser, error)
func (*EnglishTokeniser) GetTerms ¶
func (e *EnglishTokeniser) GetTerms(text string, operation func(string))
func (*EnglishTokeniser) IsStopWord ¶
func (e *EnglishTokeniser) IsStopWord(word string) bool
func (*EnglishTokeniser) Normalise ¶
func (e *EnglishTokeniser) Normalise(token string) string
func (*EnglishTokeniser) NormaliseMany ¶
func (e *EnglishTokeniser) NormaliseMany(tokens []string) []string
func (*EnglishTokeniser) Tokenise ¶
func (e *EnglishTokeniser) Tokenise(text string) []string
Click to show internal directories.
Click to hide internal directories.