nltk: NLTK -- the Natural Language Toolkit -- is a suite of open source
Python modules, data sets and tutorials supporting research and
development in natural language processing.
nltk.classify: Classes and interfaces for labeling tokens with category labels (or
class
labels).
nltk.classify.api: Interfaces for labeling tokens with category labels (or class
labels).
nltk.classify.decisiontree: A classifier model that decides which label to assign to a token on
the basis of a tree structure, where branches correspond to
conditions on feature values, and leaves correspond to label
assignments.
nltk.classify.mallet: A set of functions used to interface with the external Mallet machine
learning package.
nltk.classify.maxent: A classifier model based on maximum entropy modeling framework.
nltk.classify.megam: A set of functions used to interface with the external megam
maxent optimization package.
nltk.corpus.chat80: Chat-80 was a natural language system which allowed the user to
interrogate a Prolog knowledge base in the domain of world
geography.
nltk.corpus.reader.ycoe: Corpus reader for the York-Toronto-Helsinki Parsed Corpus of Old
English Prose (YCOE), a 1.5 million word syntactically-annotated
corpus of Old English prose texts.
nltk.data: Functions to find and load NLTK resource
files, such as corpora, grammars, and saved processing objects.
nltk.decorators: Decorator module by Michele Simionato <michelesimionato@libero.it>
Copyright Michele Simionato, distributed under the terms of the BSD License (see below).
nltk.detect: Functions for detecting a token's features.
nltk.draw: Tools for graphically displaying and interacting with the objects
and processing classes defined by the Toolkit.
nltk.misc.sort: This module provides a variety of list sorting algorithms, to
illustrate the many different algorithms (recipes) for solving a
problem, and how to analyze algorithms experimentally.
nltk.sem: This package contains classes for representing semantic structure
in formulas of first-order logic and for evaluating such formulas
in set-theoretic models.
nltk.sem.evaluate: This module provides data structures for representing first-order
models.
nltk.sem.logic: A version of first order predicate logic, built on top of the
untyped lambda calculus.
nltk.sem.relextract: Code for extracting relational triples from the ieer and conll2002
corpora.
nltk.sem.util: Utility functions for batch-processing sentences: parsing and
extraction of the semantic representation of the root node of the
the syntax tree, followed by evaluation of the semantic
representation in a first-order model.
nltk.stem: Interfaces used to remove morphological affixes from words, leaving
only the word stem.
nltk.tag.crf: An interface to Mallet's Linear Chain Conditional Random Field
(LC-CRF) implementation.
nltk.tag.hmm: Hidden Markov Models (HMMs) largely used to assign the correct
label sequence to sequential data or assess the probability of a
given label and data sequence.
nltk.tag.sequential: Classes for tagging sentences sequentially, left to right.
nltk.tokenize.regexp: Tokenizers that divide strings into substrings using regular
expressions that can match either tokens or separators between
tokens.
nltk.wordnet.browse: Natural Language Toolkit: Wordnet Interface: Wordnet Text Mode
Browser See also the NLTK Wordnet Graphical Browser in
nltk_contrib.wordnet
nltk.wordnet.browser.browserver: BrowServer is a server for browsing the NLTK Wordnet database It
first launches a browser client to be used for browsing and then
starts serving the requests of that and maybe other clients