TY - CONF
TI - Algebraic classifiers: a generic approach to fast cross-validation, online training, and parallel training
AU - Izbicki, Michael
AB - We use abstract algebra to derive new algorithms for fast cross-validation, online learning, and parallel learning. To use these algorithms on a classification model, we must show that the model has appropriate algebraic structure. It is easy to give algebraic structure to some models, and we do this explicitly for Bayesian classifiers and a novel variation of decision stumps called HomStumps. But not all classifiers have an obvious structure, so we introduce the Free HomTrainer. This can be used to give a "generic" algebraic structure to any classifier. We use the Free HomTrainer to give algebraic structure to bagging and boosting. In so doing, we derive novel online and parallel algorithms, and present the first fast cross-validation schemes for these classifiers.
C3 - ICML
DA - 2013///
PY - 2013
DP - Semantic Scholar
ST - Algebraic classifiers
KW - Algebra
KW - Categorical ML
KW - Machine learning
ER -
TY - COMP
TI - dmurfet/2simplicialtransformer
AU - Murfet, Daniel
AB - Code for the 2-simplicial Transformer paper. Contribute to dmurfet/2simplicialtransformer development by creating an account on GitHub.
DA - 2019/10/14/T08:10:47Z
PY - 2019
DP - GitHub
LA - Python
UR - https://github.com/dmurfet/2simplicialtransformer
Y2 - 2019/11/22/16:50:05
KW - Abstract machines
KW - Algebra
KW - Implementation
KW - Machine learning
KW - Semantics
ER -
TY - JOUR
TI - Logic and the $2$-Simplicial Transformer
AU - Murfet, Daniel
AU - Clift, James
AU - Doryn, Dmitry
AU - Wallbridge, James
T2 - arXiv:1909.00668 [cs, stat]
AB - We introduce the $2$-simplicial Transformer, an extension of the Transformer which includes a form of higher-dimensional attention generalising the dot-product attention, and uses this attention to update entity representations with tensor products of value vectors. We show that this architecture is a useful inductive bias for logical reasoning in the context of deep reinforcement learning.
DA - 2019/09/02/
PY - 2019
DP - arXiv.org
UR - http://arxiv.org/abs/1909.00668
Y2 - 2019/11/21/20:31:14
KW - Abstract machines
KW - Algebra
KW - Machine learning
KW - Semantics
ER -
TY - JOUR
TI - Is There Something Out There? Inferring Space from Sensorimotor Dependencies
AU - Philipona, David
AU - O’Regan, J.
AU - Nadal, Jean-Pierre
T2 - Neural computation
AB - This letter suggests that in biological organisms, the perceived structure of reality, in particular the notions of body, environment, space, object, and attribute, could be a consequence of an effort on the part of brains to account for the dependency between their inputs and their outputs in terms of a small number of parameters. To validate this idea, a procedure is demonstrated whereby the brain of a (simulated) organism with arbitrary input and output connectivity can deduce the dimensionality of the rigid group of the space underlying its input-output relationship, that is, the dimension of what the organism will call physical space.
DA - 2003/10/01/
PY - 2003
DO - 10/frg7gs
DP - ResearchGate
VL - 15
SP - 2029
EP - 49
J2 - Neural computation
ST - Is There Something Out There?
KW - Algebra
KW - Neuroscience
ER -
TY - ELEC
TI - Algebraic Geometry and Statistical Learning Theory
AU - Watanabe, Sumio
T2 - Cambridge Core
AB - Cambridge Core - Pattern Recognition and Machine Learning - Algebraic Geometry and Statistical Learning Theory - by Sumio Watanabe
DA - 2009/08//
PY - 2009
LA - en
UR - /core/books/algebraic-geometry-and-statistical-learning-theory/9C8FD1BDC817E2FC79117C7F41544A3A
Y2 - 2019/11/22/18:05:57
KW - Algebra
KW - Bayesianism
KW - Purely theoretical
KW - Statistical learning theory
ER -