Abstract
How smart can a micron-sized bag of chemicals be? How can an artificial or real cell make inferences about its environment? From which kinds of probability distributions can chemical reaction networks sample? We begin tackling these questions by showing three ways in which a stochastic chemical reaction network can implement a Boltzmann machine, a stochastic neural network model that can generate a wide range of probability distributions and compute conditional probabilities. The resulting models, and the associated theorems, provide a road map for constructing chemical reaction networks that exploit their native stochasticity as a computational resource. Finally, to show the potential of our models, we simulate a chemical Boltzmann machine to classify and generate MNIST digits in-silico.
W. Poole, A. Ortiz-Muñoz and A. Behera—Contributed Equally.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bray, D.: Protein molecules as computational elements in living cells. Nature 376(6538), 307 (1995)
Bray, D.: Wetware: A Computer in Every Living Cell. Yale University Press, New Haven (2009)
McAdams, H.H., Arkin, A.: Stochastic mechanisms in gene expression. Proc. Natl. Acad. Sci. 94(3), 814–819 (1997)
Elowitz, M.B., Levine, A.J., Siggia, E.D., Swain, P.S.: Stochastic gene expression in a single cell. Science 297(5584), 1183–1186 (2002)
Perkins, T.J., Swain, P.S.: Strategies for cellular decision making. Mol. Syst. Biol. 5(1), 326 (2009)
Muroga, S.: Threshold Logic and Its Applications. Wiley Interscience, New York (1971)
Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. 79(8), 2554–2558 (1982)
Hinton, G.E., Sejnowski, T.J., Ackley, D.H.: Boltzmann Machines: Constraint Satisfaction Networks that Learn. Department of Computer Science, Carnegie-Mellon University, Pittsburgh (1984)
Bray, D.: Intracellular signalling as a parallel distributed process. J. Theor. Biol. 143(2), 215–231 (1990)
Hellingwerf, K.J., Postma, P.W., Tommassen, J., Westerhoff, H.V.: Signal transduction in bacteria: phospho-neural network(s) in Escherichia coli. FEMS Microbiol. Rev. 16(4), 309–321 (1995)
Mjolsness, E., Sharp, D.H., Reinitz, J.: A connectionist model of development. J. Theor. Biol. 152(4), 429–453 (1991)
Mestl, T., Lemay, C., Glass, L.: Chaos in high-dimensional neural and gene networks. Physica D: Nonlin. Phenom. 98(1), 33–52 (1996)
Buchler, N.E., Gerland, U., Hwa, T.: On schemes of combinatorial transcription logic. Proc. Natl. Acad. Sci. 100(9), 5136–5141 (2003)
Deutsch, J.M.: Collective regulation by non-coding RNA. arXiv preprint arXiv:1409.1899 (2014)
Deutsch, J.M.: Associative memory by collective regulation of non-coding RNA. arXiv preprint arXiv:1608.05494 (2016)
Hjelmfelt, A., Weinberger, E.D., Ross, J.: Chemical implementation of neural networks and turing machines. Proc. Natl. Acad. Sci. 88(24), 10983–10987 (1991)
Hjelmfelt, A., Ross, J.: Chemical implementation and thermodynamics of collective neural networks. Proc. Natl. Acad. Sci. 89(1), 388–391 (1992)
Kim, J., Hopfield, J.J., Winfree, E.: Neural network computation by in vitro transcriptional circuits. In: Advances in Neural Information Processing Systems (NIPS), pp. 681–688 (2004)
Napp, N.E., Adams, R.P.: Message passing inference with chemical reaction networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 2247–2255 (2013)
Gopalkrishnan, M.: A scheme for molecular computation of maximum likelihood estimators for log-linear models. In: Rondelez, Y., Woods, D. (eds.) DNA 2016. LNCS, vol. 9818, pp. 3–18. Springer, Cham (2016). doi:10.1007/978-3-319-43994-5_1
Hjelmfelt, A., Schneider, F.W., Ross, J.: Pattern recognition in coupled chemical kinetic systems. Science 260, 335–335 (1993)
Kim, J., White, K.S., Winfree, E.: Construction of an in vitro bistable circuit from synthetic transcriptional switches. Mol. Syst. Biol. 2, 68 (2006)
Kim, J., Winfree, E.: Synthetic in vitro transcriptional oscillators. Mol. Syst. Biol. 7, 465 (2011)
Qian, L., Winfree, E., Bruck, J.: Neural network computation with DNA strand displacement cascades. Nature 475(7356), 368–372 (2011)
Lestas, I., Paulsson, J., Ross, N.E., Vinnicombe, G.: Noise in gene regulatory networks. IEEE Trans. Autom. Control 53, 189–200 (2008)
Lestas, I., Vinnicombe, G., Paulsson, J.: Fundamental limits on the suppression of molecular fluctuations. Nature 467(7312), 174–178 (2010)
Veening, J.W., Smits, W.K., Kuipers, O.P.: Bistability, epigenetics, and bet-hedging in bacteria. Annu. Rev. Microbiol. 62, 193–210 (2008)
Balázsi, G., van Oudenaarden, A., Collins, J.J.: Cellular decision making and biological noise: from microbes to mammals. Cell 144(6), 910–925 (2011)
Tsimring, L.S.: Noise in biology. Rep. Prog. Phys. 77(2), 26601 (2014)
Eldar, A., Elowitz, M.B.: Functional roles for noise in genetic circuits. Nature 467(7312), 167–173 (2010)
Mansinghka, V.K.: Natively probabilistic computation. Ph.D. thesis, Massachusetts Institute of Technology (2009)
Wang, S., Zhang, X., Li, Y., Bashizade, R., Yang, S., Dwyer, C., Lebeck, A.R.: Accelerating Markov random field inference using molecular optical Gibbs sampling units. In: Proceedings of the 43rd International Symposium on Computer Architecture, pp. 558–569. IEEE Press (2016)
Fiser, J., Berkes, P., Orbán, G., Lengyel, M.: Statistically optimal perception and learning: from behavior to neural representations. Trends Cogn. Sci. 14(3), 119–130 (2010)
Pouget, A., Beck, J.M., Ma, W.J., Latham, P.E.: Probabilistic brains: knowns and unknowns. Nat. Neurosci. 16(9), 1170–1178 (2013)
Ackley, D.H., Hinton, G.E., Sejnowski, T.J.: A learning algorithm for Boltzmann machines. Cogn. Sci. 9(1), 147–169 (1985)
Tanaka, T.: Mean-field theory of Boltzmann machine learning. Phys. Rev. E 58(2), 2302 (1998)
Tang, Y., Sutskever, I.: Data normalization in the learning of restricted Boltzmann machines. Department of Computer Science, University of Toronto, Technical report UTML-TR-11-2 (2011)
Taylor, G.W., Hinton, G.E.: Factored conditional restricted Boltzmann machines for modeling motion style. In: Proceedings of the 26th Annual International Conference on Machine Learning (ICML), pp. 1025–1032. ACM (2009)
Casella, G., George, E.I.: Explaining the Gibbs sampler. Am. Stat. 46(3), 167–174 (1992)
Gillespie, D.T.: Stochastic simulation of chemical kinetics. Annu. Rev. Phys. Chem. 58, 35–55 (2007)
Qian, H.: Phosphorylation energy hypothesis: open chemical systems and their biological functions. Annu. Rev. Phys. Chem. 58, 113–142 (2007)
Beard, D.A., Qian, H.: Chemical Biophysics: Quantitative Analysis of Cellular Systems. Cambridge University Press, Cambridge (2008)
Ouldridge, T.E.: The importance of thermodynamics for molecular systems, the importance of molecular systems for thermodynamics. arXiv preprint arXiv:1702.00360 (2017)
Joshi, B.: A detailed balanced reaction network is sufficient but not necessary for its Markov chain to be detailed balanced. arXiv preprint arXiv:1312.4196 (2013)
Erez, A., Byrd, T.A., Vogel, R.M., Altan-Bonnet, G., Mugler, A.: Criticality of biochemical feedback. arXiv preprint arXiv:1703.04194 (2017)
Anderson, D.F., Craciun, G., Kurtz, T.G.: Product-form stationary distributions for deficiency zero chemical reaction networks. Bull. Math. Biol. 72(8), 1947–1970 (2010)
LeCun, Y., Cortes, C., Burges, C.J.C.: The MNIST database of handwritten digits (1998)
Acknowledgements
This work was supported in part by U.S. National Science Foundation (NSF) graduate fellowships to WP and to AOM, by NSF grant CCF-1317694 to EW, and by the Gordon and Betty Moore Foundation through Grant GBMF2809 to the Caltech Programmable Molecular Technology Initiative (PMTI), by a Royal Society University Research Fellowship to TEO, and by a Bharti Centre for Communication in IIT Bombay award to AB.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
A Appendix
A Appendix
1.1 A.1 Application of Theorem 2: The Direct CBM Must Use Implicit Fuel Species
Here, we use Theorem 2 to analyze the direct implementation of a CBM and show that it cannot be detailed balanced and thereby requires implicit fuel molecules. First, notice that the the conservation laws used in this construction are of a simple form. The states accessible by \((X_i^{ON},X_i^{OFF})\) are independent of \((X_j^{ON},X_j^{OFF})\) for \(i \ne j\), and therefore the reachability class is a product over the subspaces of each individual node. As a consequence, by Theorem 2, the system must be out of equilibrium and violate detailed balance at the level of the CRN because, by construction, this system is equivalent to a BM and has correlations between nodes i and j whenever \(w_{ij} \ne 0\). In physical terms, the presence of catalysts cannot influence the equilibrium yield of a species, and therefore a circuit which uses catalysis to bias distributions of species must be powered by a supply of chemical fuel molecules [41,42,43]. It is also worth noting that, as a consequence, this scheme cannot be implemented by tuning of (free) energies; it is fundamentally necessary to carefully tune all of the rate constants individually (via implicit fuel molecules) to ensure that detailed balance is maintained at the level of the Markov chain for the species of interest.
1.2 A.2 BM Training and TCBM Simulation Details
We trained a BM using stochastic gradient descent on the MNIST dataset, down sampled to be 10 pixels by 10 pixels [47]. The BM has 100 visible image units (representing a 10\(\,\times \,\)10 image), 10 visible class nodes, and 40 hidden nodes as depicted in Fig. 4B. Our training data consisted of the concatenation of down sampled MNIST images and their classes projected onto the 10 class nodes. The weights and biases of the trained BM were converted to reaction rates for a CBM using the Taylor series approximation. This CBM consists of 300 species, 300 unimolecular reactions and 22350 bimolecular reactions. The resulting CBM was then compared side-by-side with the trained BM on image classification and generation. The BM was simulated using custom Gibbs sampling written in Python. The CRN was simulated on a custom Stochastic Simulation Algorithm (SSA) [40] algorithm written in Cython. All simulations, including network training, were run locally on a notebook or on a single high performance Amazon Cloud server.
Classification was carried out on all 10000 MNIST validation images using both the BM and the CBM. Each 10 by 10 gray-scale image was converted to a binary sample image by comparing the gray-scale image’s pixels (which are represented as real numbers between 0 and 1) to a uniform distribution over the same range. The network’s image units were then clamped to the binary sample and the hidden units and class units were allowed to reach steady state. This process was carried out 3 times for each MNIST validation image, resulting in 30000 sample images being classified. Raw classification scores were computed by averaging the class nodes’ outputs for 20000 simulation steps after 20000 steps of burn-in (Gibbs sampling for the BM, SSA for the CBM). Max classification was computed by taking the most probable class from the raw classification output. Raw classification and max classification confusion heatmaps, showing the average classification across all sample images as a function of the true label are shown in Fig. 4 panels C and D for a BM and in Fig. 4 panels F and G for a CBM.
Image generation was carried out by clamping the class nodes with a single class, 0...9, taking the value of 1 and all other classes being 0, and then allowing the network to reach steady state. Generated images were computed by averaging the image nodes over 50000 simulation steps (Gibbs sampling for the BM, SSA for the CBM) after 25000 steps of burn-in. Generation results are shown in Fig. 4E for a BM and Fig. 4H for a CBM.
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Poole, W. et al. (2017). Chemical Boltzmann Machines. In: Brijder, R., Qian, L. (eds) DNA Computing and Molecular Programming. DNA 2017. Lecture Notes in Computer Science(), vol 10467. Springer, Cham. https://doi.org/10.1007/978-3-319-66799-7_14
Download citation
DOI: https://doi.org/10.1007/978-3-319-66799-7_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-66798-0
Online ISBN: 978-3-319-66799-7
eBook Packages: Computer ScienceComputer Science (R0)