Skip to main content

Chemical Boltzmann Machines

  • Conference paper
  • First Online:
DNA Computing and Molecular Programming (DNA 2017)

Abstract

How smart can a micron-sized bag of chemicals be? How can an artificial or real cell make inferences about its environment? From which kinds of probability distributions can chemical reaction networks sample? We begin tackling these questions by showing three ways in which a stochastic chemical reaction network can implement a Boltzmann machine, a stochastic neural network model that can generate a wide range of probability distributions and compute conditional probabilities. The resulting models, and the associated theorems, provide a road map for constructing chemical reaction networks that exploit their native stochasticity as a computational resource. Finally, to show the potential of our models, we simulate a chemical Boltzmann machine to classify and generate MNIST digits in-silico.

W. Poole, A. Ortiz-Muñoz and A. Behera—Contributed Equally.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bray, D.: Protein molecules as computational elements in living cells. Nature 376(6538), 307 (1995)

    Article  Google Scholar 

  2. Bray, D.: Wetware: A Computer in Every Living Cell. Yale University Press, New Haven (2009)

    Google Scholar 

  3. McAdams, H.H., Arkin, A.: Stochastic mechanisms in gene expression. Proc. Natl. Acad. Sci. 94(3), 814–819 (1997)

    Article  Google Scholar 

  4. Elowitz, M.B., Levine, A.J., Siggia, E.D., Swain, P.S.: Stochastic gene expression in a single cell. Science 297(5584), 1183–1186 (2002)

    Article  Google Scholar 

  5. Perkins, T.J., Swain, P.S.: Strategies for cellular decision making. Mol. Syst. Biol. 5(1), 326 (2009)

    Google Scholar 

  6. Muroga, S.: Threshold Logic and Its Applications. Wiley Interscience, New York (1971)

    MATH  Google Scholar 

  7. Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. 79(8), 2554–2558 (1982)

    Article  MathSciNet  Google Scholar 

  8. Hinton, G.E., Sejnowski, T.J., Ackley, D.H.: Boltzmann Machines: Constraint Satisfaction Networks that Learn. Department of Computer Science, Carnegie-Mellon University, Pittsburgh (1984)

    Google Scholar 

  9. Bray, D.: Intracellular signalling as a parallel distributed process. J. Theor. Biol. 143(2), 215–231 (1990)

    Article  Google Scholar 

  10. Hellingwerf, K.J., Postma, P.W., Tommassen, J., Westerhoff, H.V.: Signal transduction in bacteria: phospho-neural network(s) in Escherichia coli. FEMS Microbiol. Rev. 16(4), 309–321 (1995)

    Article  Google Scholar 

  11. Mjolsness, E., Sharp, D.H., Reinitz, J.: A connectionist model of development. J. Theor. Biol. 152(4), 429–453 (1991)

    Article  Google Scholar 

  12. Mestl, T., Lemay, C., Glass, L.: Chaos in high-dimensional neural and gene networks. Physica D: Nonlin. Phenom. 98(1), 33–52 (1996)

    Article  MathSciNet  MATH  Google Scholar 

  13. Buchler, N.E., Gerland, U., Hwa, T.: On schemes of combinatorial transcription logic. Proc. Natl. Acad. Sci. 100(9), 5136–5141 (2003)

    Article  Google Scholar 

  14. Deutsch, J.M.: Collective regulation by non-coding RNA. arXiv preprint arXiv:1409.1899 (2014)

  15. Deutsch, J.M.: Associative memory by collective regulation of non-coding RNA. arXiv preprint arXiv:1608.05494 (2016)

  16. Hjelmfelt, A., Weinberger, E.D., Ross, J.: Chemical implementation of neural networks and turing machines. Proc. Natl. Acad. Sci. 88(24), 10983–10987 (1991)

    Article  MATH  Google Scholar 

  17. Hjelmfelt, A., Ross, J.: Chemical implementation and thermodynamics of collective neural networks. Proc. Natl. Acad. Sci. 89(1), 388–391 (1992)

    Article  Google Scholar 

  18. Kim, J., Hopfield, J.J., Winfree, E.: Neural network computation by in vitro transcriptional circuits. In: Advances in Neural Information Processing Systems (NIPS), pp. 681–688 (2004)

    Google Scholar 

  19. Napp, N.E., Adams, R.P.: Message passing inference with chemical reaction networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 2247–2255 (2013)

    Google Scholar 

  20. Gopalkrishnan, M.: A scheme for molecular computation of maximum likelihood estimators for log-linear models. In: Rondelez, Y., Woods, D. (eds.) DNA 2016. LNCS, vol. 9818, pp. 3–18. Springer, Cham (2016). doi:10.1007/978-3-319-43994-5_1

    Chapter  Google Scholar 

  21. Hjelmfelt, A., Schneider, F.W., Ross, J.: Pattern recognition in coupled chemical kinetic systems. Science 260, 335–335 (1993)

    Article  Google Scholar 

  22. Kim, J., White, K.S., Winfree, E.: Construction of an in vitro bistable circuit from synthetic transcriptional switches. Mol. Syst. Biol. 2, 68 (2006)

    Article  Google Scholar 

  23. Kim, J., Winfree, E.: Synthetic in vitro transcriptional oscillators. Mol. Syst. Biol. 7, 465 (2011)

    Article  Google Scholar 

  24. Qian, L., Winfree, E., Bruck, J.: Neural network computation with DNA strand displacement cascades. Nature 475(7356), 368–372 (2011)

    Article  Google Scholar 

  25. Lestas, I., Paulsson, J., Ross, N.E., Vinnicombe, G.: Noise in gene regulatory networks. IEEE Trans. Autom. Control 53, 189–200 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  26. Lestas, I., Vinnicombe, G., Paulsson, J.: Fundamental limits on the suppression of molecular fluctuations. Nature 467(7312), 174–178 (2010)

    Article  Google Scholar 

  27. Veening, J.W., Smits, W.K., Kuipers, O.P.: Bistability, epigenetics, and bet-hedging in bacteria. Annu. Rev. Microbiol. 62, 193–210 (2008)

    Article  Google Scholar 

  28. Balázsi, G., van Oudenaarden, A., Collins, J.J.: Cellular decision making and biological noise: from microbes to mammals. Cell 144(6), 910–925 (2011)

    Article  Google Scholar 

  29. Tsimring, L.S.: Noise in biology. Rep. Prog. Phys. 77(2), 26601 (2014)

    Article  Google Scholar 

  30. Eldar, A., Elowitz, M.B.: Functional roles for noise in genetic circuits. Nature 467(7312), 167–173 (2010)

    Article  Google Scholar 

  31. Mansinghka, V.K.: Natively probabilistic computation. Ph.D. thesis, Massachusetts Institute of Technology (2009)

    Google Scholar 

  32. Wang, S., Zhang, X., Li, Y., Bashizade, R., Yang, S., Dwyer, C., Lebeck, A.R.: Accelerating Markov random field inference using molecular optical Gibbs sampling units. In: Proceedings of the 43rd International Symposium on Computer Architecture, pp. 558–569. IEEE Press (2016)

    Google Scholar 

  33. Fiser, J., Berkes, P., Orbán, G., Lengyel, M.: Statistically optimal perception and learning: from behavior to neural representations. Trends Cogn. Sci. 14(3), 119–130 (2010)

    Article  Google Scholar 

  34. Pouget, A., Beck, J.M., Ma, W.J., Latham, P.E.: Probabilistic brains: knowns and unknowns. Nat. Neurosci. 16(9), 1170–1178 (2013)

    Article  Google Scholar 

  35. Ackley, D.H., Hinton, G.E., Sejnowski, T.J.: A learning algorithm for Boltzmann machines. Cogn. Sci. 9(1), 147–169 (1985)

    Article  Google Scholar 

  36. Tanaka, T.: Mean-field theory of Boltzmann machine learning. Phys. Rev. E 58(2), 2302 (1998)

    Article  Google Scholar 

  37. Tang, Y., Sutskever, I.: Data normalization in the learning of restricted Boltzmann machines. Department of Computer Science, University of Toronto, Technical report UTML-TR-11-2 (2011)

    Google Scholar 

  38. Taylor, G.W., Hinton, G.E.: Factored conditional restricted Boltzmann machines for modeling motion style. In: Proceedings of the 26th Annual International Conference on Machine Learning (ICML), pp. 1025–1032. ACM (2009)

    Google Scholar 

  39. Casella, G., George, E.I.: Explaining the Gibbs sampler. Am. Stat. 46(3), 167–174 (1992)

    MathSciNet  Google Scholar 

  40. Gillespie, D.T.: Stochastic simulation of chemical kinetics. Annu. Rev. Phys. Chem. 58, 35–55 (2007)

    Article  Google Scholar 

  41. Qian, H.: Phosphorylation energy hypothesis: open chemical systems and their biological functions. Annu. Rev. Phys. Chem. 58, 113–142 (2007)

    Article  Google Scholar 

  42. Beard, D.A., Qian, H.: Chemical Biophysics: Quantitative Analysis of Cellular Systems. Cambridge University Press, Cambridge (2008)

    Book  MATH  Google Scholar 

  43. Ouldridge, T.E.: The importance of thermodynamics for molecular systems, the importance of molecular systems for thermodynamics. arXiv preprint arXiv:1702.00360 (2017)

  44. Joshi, B.: A detailed balanced reaction network is sufficient but not necessary for its Markov chain to be detailed balanced. arXiv preprint arXiv:1312.4196 (2013)

  45. Erez, A., Byrd, T.A., Vogel, R.M., Altan-Bonnet, G., Mugler, A.: Criticality of biochemical feedback. arXiv preprint arXiv:1703.04194 (2017)

  46. Anderson, D.F., Craciun, G., Kurtz, T.G.: Product-form stationary distributions for deficiency zero chemical reaction networks. Bull. Math. Biol. 72(8), 1947–1970 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  47. LeCun, Y., Cortes, C., Burges, C.J.C.: The MNIST database of handwritten digits (1998)

    Google Scholar 

Download references

Acknowledgements

This work was supported in part by U.S. National Science Foundation (NSF) graduate fellowships to WP and to AOM, by NSF grant CCF-1317694 to EW, and by the Gordon and Betty Moore Foundation through Grant GBMF2809 to the Caltech Programmable Molecular Technology Initiative (PMTI), by a Royal Society University Research Fellowship to TEO, and by a Bharti Centre for Communication in IIT Bombay award to AB.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Manoj Gopalkrishnan .

Editor information

Editors and Affiliations

A Appendix

A Appendix

1.1 A.1 Application of Theorem 2: The Direct CBM Must Use Implicit Fuel Species

Here, we use Theorem 2 to analyze the direct implementation of a CBM and show that it cannot be detailed balanced and thereby requires implicit fuel molecules. First, notice that the the conservation laws used in this construction are of a simple form. The states accessible by \((X_i^{ON},X_i^{OFF})\) are independent of \((X_j^{ON},X_j^{OFF})\) for \(i \ne j\), and therefore the reachability class is a product over the subspaces of each individual node. As a consequence, by Theorem 2, the system must be out of equilibrium and violate detailed balance at the level of the CRN because, by construction, this system is equivalent to a BM and has correlations between nodes i and j whenever \(w_{ij} \ne 0\). In physical terms, the presence of catalysts cannot influence the equilibrium yield of a species, and therefore a circuit which uses catalysis to bias distributions of species must be powered by a supply of chemical fuel molecules [41,42,43]. It is also worth noting that, as a consequence, this scheme cannot be implemented by tuning of (free) energies; it is fundamentally necessary to carefully tune all of the rate constants individually (via implicit fuel molecules) to ensure that detailed balance is maintained at the level of the Markov chain for the species of interest.

1.2 A.2 BM Training and TCBM Simulation Details

We trained a BM using stochastic gradient descent on the MNIST dataset, down sampled to be 10 pixels by 10 pixels [47]. The BM has 100 visible image units (representing a 10\(\,\times \,\)10 image), 10 visible class nodes, and 40 hidden nodes as depicted in Fig. 4B. Our training data consisted of the concatenation of down sampled MNIST images and their classes projected onto the 10 class nodes. The weights and biases of the trained BM were converted to reaction rates for a CBM using the Taylor series approximation. This CBM consists of 300 species, 300 unimolecular reactions and 22350 bimolecular reactions. The resulting CBM was then compared side-by-side with the trained BM on image classification and generation. The BM was simulated using custom Gibbs sampling written in Python. The CRN was simulated on a custom Stochastic Simulation Algorithm (SSA) [40] algorithm written in Cython. All simulations, including network training, were run locally on a notebook or on a single high performance Amazon Cloud server.

Classification was carried out on all 10000 MNIST validation images using both the BM and the CBM. Each 10 by 10 gray-scale image was converted to a binary sample image by comparing the gray-scale image’s pixels (which are represented as real numbers between 0 and 1) to a uniform distribution over the same range. The network’s image units were then clamped to the binary sample and the hidden units and class units were allowed to reach steady state. This process was carried out 3 times for each MNIST validation image, resulting in 30000 sample images being classified. Raw classification scores were computed by averaging the class nodes’ outputs for 20000 simulation steps after 20000 steps of burn-in (Gibbs sampling for the BM, SSA for the CBM). Max classification was computed by taking the most probable class from the raw classification output. Raw classification and max classification confusion heatmaps, showing the average classification across all sample images as a function of the true label are shown in Fig. 4 panels C and D for a BM and in Fig. 4 panels F and G for a CBM.

Image generation was carried out by clamping the class nodes with a single class, 0...9, taking the value of 1 and all other classes being 0, and then allowing the network to reach steady state. Generated images were computed by averaging the image nodes over 50000 simulation steps (Gibbs sampling for the BM, SSA for the CBM) after 25000 steps of burn-in. Generation results are shown in Fig. 4E for a BM and Fig. 4H for a CBM.

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Poole, W. et al. (2017). Chemical Boltzmann Machines. In: Brijder, R., Qian, L. (eds) DNA Computing and Molecular Programming. DNA 2017. Lecture Notes in Computer Science(), vol 10467. Springer, Cham. https://doi.org/10.1007/978-3-319-66799-7_14

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-66799-7_14

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-66798-0

  • Online ISBN: 978-3-319-66799-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics