@misc{borchert_amzn/milan_2019,
title = {amzn/milan},
copyright = {Apache-2.0},
url = {https://github.com/amzn/milan},
abstract = {Milan is a Scala API and runtime infrastructure for building data-oriented systems, built on top of Apache Flink.},
urldate = {2019-11-27},
publisher = {Amazon},
author = {Borchert, Tom},
month = nov,
year = {2019},
note = {ZSCC: NoCitationData[s0]
original-date: 2019-11-06T22:37:51Z},
keywords = {Implementation, Machine learning, Probabilistic programming}
}
@book{engeler_combinatory_1995,
series = {Progress in {Theoretical} {Computer} {Science}},
title = {The {Combinatory} {Programme}},
isbn = {978-0-8176-3801-6},
url = {https://www.springer.com/gb/book/9780817638016},
abstract = {Combinatory logic started as a programme in the foundation of mathematics and in an historical context at a time when such endeavours attracted the most gifted among the mathematicians. This small volume arose under quite differÂ ent circumstances, namely within the context of reworking the mathematical foundations of computer science. I have been very lucky in finding gifted students who agreed to work with me and chose, for their Ph. D. theses, subjects that arose from my own attempts 1 to create a coherent mathematical view of these foundations. The result of this collaborative work is presented here in the hope that it does justice to the individual contributor and that the reader has a chance of judging the work as a whole. E. Engeler ETH Zurich, April 1994 lCollected in Chapter III, An Algebraization of Algorithmics, in Algorithmic Properties of Structures, Selected Papers of Erwin Engeler, World Scientific PubJ. Co. , Singapore, 1993, pp. 183-257. I Historical and Philosophical Background Erwin Engeler In the fall of 1928 a young American turned up at the Mathematical Institute of Gottingen, a mecca of mathematicians at the time; he was a young man with a dream and his name was H. B. Curry. He felt that he had the tools in hand with which to solve the problem of foundations of mathematics mice and for all. His was an approach that came to be called "formalist" and embodied that later became known as Combinatory Logic.},
language = {en},
urldate = {2019-11-26},
publisher = {BirkhĂ¤user Basel},
author = {Engeler, Erwin},
year = {1995},
doi = {10.1007/978-1-4612-4268-0},
note = {ZSCC: NoCitationData[s1] },
keywords = {Algebra, Programming language theory, Purely theoretical}
}
@article{ghahramani_probabilistic_2015,
title = {Probabilistic machine learning and artificial intelligence},
volume = {521},
issn = {0028-0836, 1476-4687},
url = {http://www.nature.com/articles/nature14541},
doi = {10/gdxwhq},
language = {en},
number = {7553},
urldate = {2019-11-28},
journal = {Nature},
author = {Ghahramani, Zoubin},
month = may,
year = {2015},
note = {ZSCC: 0000611},
keywords = {Bayesian inference, Classical ML, Machine learning, Probabilistic programming},
pages = {452--459}
}
@article{jacobs_predicate/state_2016,
series = {The {Thirty}-second {Conference} on the {Mathematical} {Foundations} of {Programming} {Semantics} ({MFPS} {XXXII})},
title = {A {Predicate}/{State} {Transformer} {Semantics} for {Bayesian} {Learning}},
volume = {325},
issn = {1571-0661},
url = {http://www.sciencedirect.com/science/article/pii/S1571066116300883},
doi = {10/ggdgbb},
abstract = {This paper establishes a link between Bayesian inference (learning) and predicate and state transformer operations from programming semantics and logic. Specifically, a very general definition of backward inference is given via first applying a predicate transformer and then conditioning. Analogously, forward inference involves first conditioning and then applying a state transformer. These definitions are illustrated in many examples in discrete and continuous probability theory and also in quantum theory.},
language = {en},
urldate = {2019-11-24},
journal = {Electronic Notes in Theoretical Computer Science},
author = {Jacobs, Bart and Zanasi, Fabio},
month = oct,
year = {2016},
note = {ZSCC: 0000030},
keywords = {Bayesianism, Categorical ML, Categorical probability theory, Effectus theory, Programming language theory, Semantics},
pages = {185--200}
}
@misc{murfet_dmurfet/2simplicialtransformer_2019,
title = {dmurfet/2simplicialtransformer},
url = {https://github.com/dmurfet/2simplicialtransformer},
abstract = {Code for the 2-simplicial Transformer paper. Contribute to dmurfet/2simplicialtransformer development by creating an account on GitHub.},
urldate = {2019-11-22},
author = {Murfet, Daniel},
month = oct,
year = {2019},
note = {ZSCC: NoCitationData[s0]
original-date: 2019-08-29T13:26:13Z},
keywords = {Abstract machines, Algebra, Implementation, Machine learning, Semantics}
}
@misc{murfet_dmurfet/deeplinearlogic_2018,
title = {dmurfet/deeplinearlogic},
url = {https://github.com/dmurfet/deeplinearlogic},
abstract = {Deep learning and linear logic. Contribute to dmurfet/deeplinearlogic development by creating an account on GitHub.},
urldate = {2019-11-22},
author = {Murfet, Daniel},
month = jul,
year = {2018},
note = {ZSCC: NoCitationData[s0]
original-date: 2016-11-05T09:17:10Z},
keywords = {Categorical ML, Implementation, Linear logic, Machine learning, Semantics}
}
@misc{murfet_dmurfet/polysemantics_2018,
title = {dmurfet/polysemantics},
url = {https://github.com/dmurfet/polysemantics},
abstract = {Polynomial semantics of linear logic. Contribute to dmurfet/polysemantics development by creating an account on GitHub.},
urldate = {2019-11-22},
author = {Murfet, Daniel},
month = apr,
year = {2018},
note = {ZSCC: NoCitationData[s0]
original-date: 2016-02-23T03:29:42Z},
keywords = {Categorical ML, Implementation, Linear logic, Machine learning, Semantics}
}
@article{murfet_logic_2019,
title = {Logic and the \$2\$-{Simplicial} {Transformer}},
url = {http://arxiv.org/abs/1909.00668},
abstract = {We introduce the \$2\$-simplicial Transformer, an extension of the Transformer which includes a form of higher-dimensional attention generalising the dot-product attention, and uses this attention to update entity representations with tensor products of value vectors. We show that this architecture is a useful inductive bias for logical reasoning in the context of deep reinforcement learning.},
urldate = {2019-11-21},
journal = {arXiv:1909.00668 [cs, stat]},
author = {Murfet, Daniel and Clift, James and Doryn, Dmitry and Wallbridge, James},
month = sep,
year = {2019},
note = {ZSCC: 0000000
arXiv: 1909.00668
version: 1},
keywords = {Abstract machines, Algebra, Machine learning, Semantics}
}
@misc{murfet_linear_nodate,
title = {Linear logic and deep learning},
language = {en},
author = {Murfet, Daniel and Hu, Huiyi},
note = {ZSCC: NoCitationData[s0]},
keywords = {Categorical ML, Linear logic, Machine learning, Semantics}
}
@article{tran_deep_2017,
title = {Deep {Probabilistic} {Programming}},
url = {http://arxiv.org/abs/1701.03757},
abstract = {We propose Edward, a Turing-complete probabilistic programming language. Edward defines two compositional representations---random variables and inference. By treating inference as a first class citizen, on a par with modeling, we show that probabilistic programming can be as flexible and computationally efficient as traditional deep learning. For flexibility, Edward makes it easy to fit the same model using a variety of composable inference methods, ranging from point estimation to variational inference to MCMC. In addition, Edward can reuse the modeling representation as part of inference, facilitating the design of rich variational models and generative adversarial networks. For efficiency, Edward is integrated into TensorFlow, providing significant speedups over existing probabilistic systems. For example, we show on a benchmark logistic regression task that Edward is at least 35x faster than Stan and 6x faster than PyMC3. Further, Edward incurs no runtime overhead: it is as fast as handwritten TensorFlow.},
urldate = {2019-11-27},
journal = {arXiv:1701.03757 [cs, stat]},
author = {Tran, Dustin and Hoffman, Matthew D. and Saurous, Rif A. and Brevdo, Eugene and Murphy, Kevin and Blei, David M.},
month = mar,
year = {2017},
note = {ZSCC: 0000108
arXiv: 1701.03757},
keywords = {Bayesian inference, Implementation, Machine learning, Probabilistic programming}
}