Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Transduction (machine learning)
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{short description|Type of statistical inference}} {{More footnotes needed|date=April 2011}} In [[logic]], [[statistical inference]], and [[supervised learning]], '''transduction''' or '''transductive inference''' is [[reasoning]] from observed, specific (training) cases to specific (test) cases. In contrast, [[induction (philosophy)|induction]] is reasoning from observed training cases to general rules, which are then applied to the test cases. The distinction is most interesting in cases where the predictions of the transductive model are not achievable by any inductive model. Note that this is caused by transductive inference on different test sets producing mutually inconsistent predictions. Transduction was introduced in a computer science context by [[Vladimir Vapnik]] in the 1990s, motivated by his view that transduction is preferable to induction since, according to him, induction requires solving a more general problem (inferring a function) before solving a more specific problem (computing outputs for new cases): "When solving a problem of interest, do not solve a more general problem as an intermediate step. Try to get the answer that you really need but not a more general one.".<ref>{{Cite journal|last=Vapnik|first=Vladimir|date=2006|title=Estimation of Dependences Based on Empirical Data|url=https://doi.org/10.1007/0-387-34239-7|journal=Information Science and Statistics|language=en-gb|pages=477|doi=10.1007/0-387-34239-7|isbn=978-0-387-30865-4 |issn=1613-9011|url-access=subscription}}</ref> An example of learning which is not inductive would be in the case of binary classification, where the inputs tend to cluster in two groups. A large set of test inputs may help in finding the clusters, thus providing useful information about the classification labels. The same predictions would not be obtainable from a model which induces a function based only on the training cases. Some people may call this an example of the closely related [[semi-supervised learning]], since Vapnik's motivation is quite different. The most well-known example of a case-bases learning algorithm is the [[k-nearest neighbor algorithm]], which is related to transductive learning algorithms.<ref>Advances in Information Retrieval: 37th European Conference on IR Research, ECIR 2015, Vienna, Austria, March 29 - April 2, 2015. Proceedings. (2015). Deutschland: Springer International Publishing. Page 96 https://books.google.com/books?id=dbpnBwAAQBAJ&pg=PA96</ref> Another example of an algorithm in this category is the Transductive [[Support Vector Machine]] (TSVM). A third possible motivation of transduction arises through the need to approximate. If exact inference is computationally prohibitive, one may at least try to make sure that the approximations are good at the test inputs. In this case, the test inputs could come from an arbitrary distribution (not necessarily related to the distribution of the training inputs), which wouldn't be allowed in semi-supervised learning. An example of an algorithm falling in this category is the [[Bayesian Committee Machine]] (BCM).
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)