
🚨 Get your Google Knowledge Panel
◉ Displays your name, photo, and profession in Google Search
Click here to get started now
Statistical classification
{{#ifeq:categorization of data using statistics|none|Template:SHORTDESC:|
}}{{#ifeq:{{{pagetype}}}|Disambiguation pages||{{#ifeq:{{safesubst:#invoke:pagetype|main}}|exclude||{{#ifeq:{{#switch: 0 | 2 | 3 | 4 | 5 | 6 | 7 | 10 | 11 | 12 | 13 | 14 | 15 | 100 | 101 | 118 | 119 | 828 | 829 | = exclude|#default=}}|exclude||[[Category:{{safesubst:#invoke:pagetype|main}} with short description]]}}}}}}{{#invoke:Check for unknown parameters|check|unknown={{#switch:
{{#if:
| {{{demospace}}}
| {{#ifeq:|
| main
| other
}}
}}
| main = | other | #default = }}|preview=Page using Template:Short description with unknown parameter "_VALUE_"|ignoreblank=y| 1 | 2 | pagetype | bot |plural }}{{#ifexpr: {{#invoke:String|len|Categorization of data using statistics}}>100 | [[Category:{{safesubst:#invoke:pagetype|main}} with long short description]]}}{{#if:Categorization of data using statistics||}}Template:Short description/lowercasecheck{{#switch:
{{#if:
| {{{demospace}}}
| {{#ifeq:|
| main
| other
}}
}}
| main = {{#invoke:SDcat |setCat}} | other | #default = }} When classification is performed by a computer, statistical methods are normally used to develop the algorithm.
Often, the individual observations are analyzed into a set of quantifiable properties, known variously as explanatory variables or features. These properties may variously be categorical (e.g. "A", "B", "AB" or "O", for blood type), ordinal (e.g. "large", "medium" or "small"), integer-valued (e.g. the number of occurrences of a particular word in an email) or real-valued (e.g. a measurement of blood pressure). Other classifiers work by comparing observations to previous observations by means of a similarity or distance function.
An algorithm that implements classification, especially in a concrete implementation, is known as a classifier. The term "classifier" sometimes also refers to the mathematical function, implemented by a classification algorithm, that maps input data to a category.
Terminology across fields is quite varied. In statistics, where classification is often done with logistic regression or a similar procedure, the properties of observations are termed explanatory variables (or independent variables, regressors, etc.), and the categories to be predicted are known as outcomes, which are considered to be possible values of the dependent variable. In machine learning, the observations are often known as instances, the explanatory variables are termed features (grouped into a feature vector), and the possible categories to be predicted are classes. Other fields may use different terminology: e.g. in community ecology, the term "classification" normally refers to cluster analysis.
Relation to other problems
Classification and clustering are examples of the more general problem of pattern recognition, which is the assignment of some sort of output value to a given input value. Other examples are regression, which assigns a real-valued output to each input; sequence labeling, which assigns a class to each member of a sequence of values (for example, part of speech tagging, which assigns a part of speech to each word in an input sentence); parsing, which assigns a parse tree to an input sentence, describing the syntactic structure of the sentence; etc.
A common subclass of classification is probabilistic classification. Algorithms of this nature use statistical inference to find the best class for a given instance. Unlike other algorithms, which simply output a "best" class, probabilistic algorithms output a probability of the instance being a member of each of the possible classes. The best class is normally then selected as the one with the highest probability. However, such an algorithm has numerous advantages over non-probabilistic classifiers:
- It can output a confidence value associated with its choice (in general, a classifier that can do this is known as a confidence-weighted classifier).
- Correspondingly, it can abstain when its confidence of choosing any particular output is too low.
- Because of the probabilities which are generated, probabilistic classifiers can be more effectively incorporated into larger machine-learning tasks, in a way that partially or completely avoids the problem of error propagation.
Frequentist procedures
Early work on statistical classification was undertaken by Fisher,[1][2] in the context of two-group problems, leading to Fisher's linear discriminant function as the rule for assigning a group to a new observation.[3] This early work assumed that data-values within each of the two groups had a multivariate normal distribution. The extension of this same context to more than two groups has also been considered with a restriction imposed that the classification rule should be linear.[3][4] Later work for the multivariate normal distribution allowed the classifier to be nonlinear:[5] several classification rules can be derived based on different adjustments of the Mahalanobis distance, with a new observation being assigned to the group whose centre has the lowest adjusted distance from the observation.
Bayesian procedures
Unlike frequentist procedures, Bayesian classification procedures provide a natural way of taking into account any available information about the relative sizes of the different groups within the overall population.[6] Bayesian procedures tend to be computationally expensive and, in the days before Markov chain Monte Carlo computations were developed, approximations for Bayesian clustering rules were devised.[7]
Some Bayesian procedures involve the calculation of group-membership probabilities: these provide a more informative outcome than a simple attribution of a single group-label to each new observation.
Binary and multiclass classification
Classification can be thought of as two separate problems – binary classification and multiclass classification. In binary classification, a better understood task, only two classes are involved, whereas multiclass classification involves assigning an object to one of several classes.[8] Since many classification methods have been developed specifically for binary classification, multiclass classification often requires the combined use of multiple binary classifiers.
Feature vectors
{{#invoke:Labelled list hatnote|labelledList|Main article|Main articles|Main page|Main pages}}
Most algorithms describe an individual instance whose category is to be predicted using a feature vector of individual, measurable properties of the instance. Each property is termed a feature, also known in statistics as an explanatory variable (or independent variable, although features may or may not be statistically independent). Features may variously be binary (e.g. "on" or "off"); categorical (e.g. "A", "B", "AB" or "O", for blood type); ordinal (e.g. "large", "medium" or "small"); integer-valued (e.g. the number of occurrences of a particular word in an email); or real-valued (e.g. a measurement of blood pressure). If the instance is an image, the feature values might correspond to the pixels of an image; if the instance is a piece of text, the feature values might be occurrence frequencies of different words. Some algorithms work only in terms of discrete data and require that real-valued or integer-valued data be discretized into groups (e.g. less than 5, between 5 and 10, or greater than 10).
Linear classifiers
{{#invoke:Labelled list hatnote|labelledList|Main article|Main articles|Main page|Main pages}} A large number of algorithms for classification can be phrased in terms of a linear function that assigns a score to each possible category k by combining the feature vector of an instance with a vector of weights, using a dot product. The predicted category is the one with the highest score. This type of score function is known as a linear predictor function and has the following general form: <math display=block>\operatorname{score}(\mathbf{X}_i, k) = \boldsymbol\beta_k \cdot \mathbf{X}_i,</math> where Xi is the feature vector for instance i, βk is the vector of weights corresponding to category k, and score(Xi, k) is the score associated with assigning instance i to category k. In discrete choice theory, where instances represent people and categories represent choices, the score is considered the utility associated with person i choosing category k.
Algorithms with this basic setup are known as linear classifiers. What distinguishes them is the procedure for determining (training) the optimal weights/coefficients and the way that the score is interpreted.
Examples of such algorithms include
- Template:Annotated link
- Template:Annotated link
- The perceptron algorithm
- Template:Annotated link
- Template:Annotated link
Algorithms
Since no single form of classification is appropriate for all data sets, a large toolkit of classification algorithms has been developed. The most commonly used include:[9]
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
Choices between different possible algorithms are frequently made on the basis of quantitative evaluation of accuracy.
Application domains
Template:See also Classification has many applications. In some of these, it is employed as a data mining procedure, while in others more detailed statistical modeling is undertaken.
- Template:Annotated link
- Template:Annotated link identification
- Template:Annotated link
- Medical image analysis and Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Drug discovery and Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Internet Template:Annotated link
- Micro-array classification
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
Template:More footnotes needed
See also
Template:Portal Template:Colbegin
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
- Template:Annotated link
References
{{#invoke:For|For}}{{SAFESUBST:#invoke:Unsubst||date=__DATE__ |$B= {{#invoke:Message box|ambox}} }}{{#invoke:Side box|main}} Category, plural categories, may refer to:
<templatestyles src="Template:TOC_right/styles.css" />{{#if:|<templatestyles src="Template:TOC limit/styles.css" />}}
General uses
- Classification, the general act of allocating things to classes/categories
Philosophy
- Category of being
- Categories (Aristotle)
- Category (Kant)
- Categories (Peirce)
- Category (Vaisheshika)
- Stoic categories
- Category mistake
Science
- Cognitive categorization, categories in cognitive science
- Statistical classification, statistical methods used to effect classification/categorization
Mathematics
- Category (mathematics), a structure consisting of objects and arrows
- Category (topology), in the context of Baire spaces
- Lusternik–Schnirelmann category, sometimes called LS-category or simply category
- Categorical data, in statistics
Linguistics
{{#invoke:Labelled list hatnote|labelledList|Main article|Main articles|Main page|Main pages}}
- Lexical category, a part of speech such as noun, preposition, etc.
- Syntactic category, a similar concept which can also include phrasal categories
- Grammatical category, a grammatical feature such as tense, gender, etc.
Other
- Category (chess tournament)
- Objective-C categories, a computer programming concept
- Pregnancy category
- Prisoner security categories in the United Kingdom
- Weight class (boxing)
- List of software categories
- Categories (word game), a classic party game
- Saffir–Simpson hurricane wind scale, a common categorization of hurricane intensities
- Categories of New Testament manuscripts
- Network cable categories: 1, 2, 3, 4, 5/5e, 6/6a, 7/7a (F)
- A classification in a system of compartmentalization (information security)
See also
- Template:Intitle
- Categorical (disambiguation)
- Category 1 (disambiguation)
- Category 2 (disambiguation)
- Category 3 (disambiguation)
- Category 4 (disambiguation)
- Category 5 (disambiguation)
- Category 6 (disambiguation)
- Category 7: The End of the World
- Category A (disambiguation)
- Category B (disambiguation)
- Category C (disambiguation)
{{#if:||Template:Disambiguation page short description}} <templatestyles src="Dmbox/styles.css" />
{{#switch:disambig
| | disambig | setindex =
| #default =
}}{{#if:||{{#ifeq:disambig|disambig|__DISAMBIG__|}}}}{{#switch:disambig | setindex =
{{#invoke:Category handler|main}}
| disambig | #default =
{{#invoke:Category handler|main}}
}}{{#switch:
{{#if:
|
| {{#ifeq:|
| main
| other
}}
}}
| main = {{safesubst:#invoke:For loop|main}}{{#if:|}}{{#if:{{safesubst:#invoke:String|match|{{#invoke:String|replace|source=Statistical classification|(disambiguation)||1}}|%s%((.-%)?)%)||-1|ignore_errors=true}}||}}
| other | #default =
}}{{#switch:
{{#if:
| {{{demospace}}}
| {{#ifeq:|
| main
| other
}}
}}
| main = {{#if:||{{#ifeq:{{#invoke:redirect|isRedirect|Talk:Statistical classification}}|yes|}}}} | other | #default = }}
- ↑ {{#invoke:Citation/CS1|citation |CitationClass=journal }}
- ↑ {{#invoke:Citation/CS1|citation |CitationClass=journal }}
- ↑ 3.0 3.1 Gnanadesikan, R. (1977) Methods for Statistical Data Analysis of Multivariate Observations, Wiley. Template:ISBN (p. 83–86)
- ↑ Rao, C.R. (1952) Advanced Statistical Methods in Multivariate Analysis, Wiley. (Section 9c)
- ↑ Anderson, T.W. (1958) An Introduction to Multivariate Statistical Analysis, Wiley.
- ↑ {{#invoke:Citation/CS1|citation |CitationClass=journal }}
- ↑ {{#invoke:Citation/CS1|citation |CitationClass=journal }}
- ↑ Har-Peled, S., Roth, D., Zimak, D. (2003) "Constraint Classification for Multiclass Classification and Ranking." In: Becker, B., Thrun, S., Obermayer, K. (Eds) Advances in Neural Information Processing Systems 15: Proceedings of the 2002 Conference, MIT Press. Template:ISBN
- ↑ {{#if: | {{{author}}} }} {{#if: https://builtin.com/data-science/tour-top-10-algorithms-machine-learning-newbies | A Tour of The Top 10 Algorithms for Machine Learning Newbies }} {{#if: | {{{publisher}}}. }} {{#if: 2019-06-10 | Accessed: 2019-06-10. }}
- Pages using short description with unknown parameters
- Pages with empty short description
- Manual of Style disambiguation pages
- Wikipedia message box parameter needs fixing
- Disambiguation pages
- Human name disambiguation pages
- Disambiguation pages with (qualified) titles
- Unsynchronized disambiguation talk pages
- Statistical classification
- Classification algorithms